According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,give...According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,given the incompleteness of the interactome and the limited knowledge of disease-associated genes,it is not obvious if the available data have sufficient coverage to map out modules associated with each disease.展开更多
According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,give...According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,given the incompleteness of the interactome and the limited knowledge of disease-associated genes,it is not obvious if the available data have sufficient coverage to map out modules associated with each disease.Here we derive mathematical conditions for the identifiability of disease modules and show that the network-based location of each disease module determines its pathobiological relationship to other diseases.For example,diseases with overlapping network modules show significant coexpression patterns,symptom similarity,and comorbidity,whereas diseases residing in separated network neighborhoods are phenotypically distinct.These tools represent an interactome-based platform to predict molecular commonalities between phenotypically related diseases,even if they do not share primary disease genes.展开更多
Optimizing routing and resource allocation in decentralized unmanned aerial vehicle(UAV)networks remains challenging due to interference and rapidly changing topologies.The authors introduce a novel framework combinin...Optimizing routing and resource allocation in decentralized unmanned aerial vehicle(UAV)networks remains challenging due to interference and rapidly changing topologies.The authors introduce a novel framework combining double deep Q-networks(DDQNs)and graph neural networks(GNNs)for joint routing and resource allocation.The framework uses GNNs to model the network topology and DDQNs to adaptively control routing and resource allocation,addressing interference and improving network performance.Simulation results show that the proposed approach outperforms traditional methods such as Closest-to-Destination(c2Dst),Max-SINR(mSINR),and Multi-Layer Perceptron(MLP)-based models,achieving approximately 23.5% improvement in throughput,50% increase in connection probability,and 17.6% reduction in number of hops,demonstrating its effectiveness in dynamic UAV networks.展开更多
The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-gener...The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-generation(5G)networks transformed mobile broadband and machine-type communications at massive scales,their properties of scaling,interference management,and latency remain a limitation in dense high mobility settings.To overcome these limitations,artificial intelligence(AI)and unmanned aerial vehicles(UAVs)have emerged as potential solutions to develop versatile,dynamic,and energy-efficient communication systems.The study proposes an AI-based UAV architecture that utilizes cooperative reinforcement learning(CoRL)to manage an autonomous network.The UAVs collaborate by sharing local observations and real-time state exchanges to optimize user connectivity,movement directions,allocate power,and resource distribution.Unlike conventional centralized or autonomous methods,CoRL involves joint state sharing and conflict-sensitive reward shaping,which ensures fair coverage,less interference,and enhanced adaptability in a dynamic urban environment.Simulations conducted in smart city scenarios with 10 UAVs and 50 ground users demonstrate that the proposed CoRL-based UAV system increases user coverage by up to 10%,achieves convergence 40%faster,and reduces latency and energy consumption by 30%compared with centralized and decentralized baselines.Furthermore,the distributed nature of the algorithm ensures scalability and flexibility,making it well-suited for future large-scale 6G deployments.The results highlighted that AI-enabled UAV systems enhance connectivity,support ultra-reliable low-latency communications(URLLC),and improve 6G network efficiency.Future work will extend the framework with adaptive modulation,beamforming-aware positioning,and real-world testbed deployment.展开更多
This paper investigates the traffic offloading optimization challenge in Space-Air-Ground Integrated Networks(SAGIN)through a novel Recursive Multi-Agent Proximal Policy Optimization(RMAPPO)algorithm.The exponential g...This paper investigates the traffic offloading optimization challenge in Space-Air-Ground Integrated Networks(SAGIN)through a novel Recursive Multi-Agent Proximal Policy Optimization(RMAPPO)algorithm.The exponential growth of mobile devices and data traffic has substantially increased network congestion,particularly in urban areas and regions with limited terrestrial infrastructure.Our approach jointly optimizes unmanned aerial vehicle(UAV)trajectories and satellite-assisted offloading strategies to simultaneously maximize data throughput,minimize energy consumption,and maintain equitable resource distribution.The proposed RMAPPO framework incorporates recurrent neural networks(RNNs)to model temporal dependencies in UAV mobility patterns and utilizes a decentralized multi-agent reinforcement learning architecture to reduce communication overhead while improving system robustness.The proposed RMAPPO algorithm was evaluated through simulation experiments,with the results indicating that it significantly enhances the cumulative traffic offloading rate of nodes and reduces the energy consumption of UAVs.展开更多
Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may r...Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.展开更多
Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and ...Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.展开更多
The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.H...The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.However,maintaining consistent forwarding states during these updates is challenging,particularly when rerouting multiple flows simultaneously.Existing approaches pay little attention to multi-flow update,where improper update sequences across data plane nodes may construct deadlock dependencies.Moreover,these methods typically involve excessive control-data plane interactions,incurring significant resource overhead and performance degradation.This paper presents P4LoF,an efficient loop-free update approach that enables the controller to reroute multiple flows through minimal interactions.P4LoF first utilizes a greedy-based algorithm to generate the shortest update dependency chain for the single-flow update.These chains are then dynamically merged into a dependency graph and resolved as a Shortest Common Super-sequence(SCS)problem to produce the update sequence of multi-flow update.To address deadlock dependencies in multi-flow updates,P4LoF builds a deadlock-fix forwarding model that leverages the flexible packet processing capabilities of the programmable data plane.Experimental results show that P4LoF reduces control-data plane interactions by at least 32.6%with modest overhead,while effectively guaranteeing loop-free consistency.展开更多
Ensuring reliable power supply in urban distribution networks is a complex and critical task.To address the increased demand during extreme scenarios,this paper proposes an optimal dispatch strategy that considers the...Ensuring reliable power supply in urban distribution networks is a complex and critical task.To address the increased demand during extreme scenarios,this paper proposes an optimal dispatch strategy that considers the coordination with virtual power plants(VPPs).The proposed strategy improves systemflexibility and responsiveness by optimizing the power adjustment of flexible resources.In the proposed strategy,theGaussian Process Regression(GPR)is firstly employed to determine the adjustable range of aggregated power within the VPP,facilitating an assessment of its potential contribution to power supply support.Then,an optimal dispatch model based on a leader-follower game is developed to maximize the benefits of the VPP and flexible resources while guaranteeing the power balance at the same time.To solve the proposed optimal dispatch model efficiently,the constraints of the problem are reformulated and resolved using the Karush-Kuhn-Tucker(KKT)optimality conditions and linear programming duality theorem.The effectiveness of the strategy is illustrated through a detailed case study.展开更多
Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the u...Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the use of virtual reality(VR)technology.VR has been demonstrated to be an effective treatment for pain associated with medical procedures,as well as for chronic pain conditions for which no effective treatment has been established.The precise mechanism by which the diversion from reality facilitated by VR contributes to the diminution of pain and anxiety has yet to be elucidated.However,the provision of positive images through VR-based visual stimulation may enhance the functionality of brain networks.The salience network is diminished,while the default mode network is enhanced.Additionally,the medial prefrontal cortex may establish a stronger connection with the default mode network,which could result in a reduction of pain and anxiety.Further research into the potential of VR technology to alleviate pain could lead to a reduction in the number of individuals who overdose on painkillers and contribute to positive change in the medical field.展开更多
Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in us...Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.展开更多
Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently...Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.展开更多
Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay di...Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.展开更多
The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and hist...The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.展开更多
With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based...With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.展开更多
For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models...For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models and physical models,each offering unique advantages but also facing limitations.Physics-informed neural networks(PINNs)provide a robust framework to integrate data-driven models with physical principles,ensuring consistency with underlying physics while enabling generalization across diverse operational conditions.This study introduces a PINN-based approach to reconstruct open circuit voltage(OCV)curves and estimate key ageing parameters at both the cell and electrode levels.These parameters include available capacity,electrode capacities,and lithium inventory capacity.The proposed method integrates OCV reconstruction models as functional components into convolutional neural networks(CNNs)and is validated using a public dataset.The results reveal that the estimated ageing parameters closely align with those obtained through offline OCV tests,with errors in reconstructed OCV curves remaining within 15 mV.This demonstrates the ability of the method to deliver fast and accurate degradation diagnostics at the electrode level,advancing the potential for precise and efficient battery health management.展开更多
Accurate estimation of mineralogy from geophysical well logs is crucial for characterizing geological formations,particularly in hydrocarbon exploration,CO_(2) sequestration,and geothermal energy development.Current t...Accurate estimation of mineralogy from geophysical well logs is crucial for characterizing geological formations,particularly in hydrocarbon exploration,CO_(2) sequestration,and geothermal energy development.Current techniques,such as multimineral petrophysical analysis,offer details into mineralogical distribution.However,it is inherently time-intensive and demands substantial geological expertise for accurate model evaluation.Furthermore,traditional machine learning techniques often struggle to predict mineralogy accurately and sometimes produce estimations that violate fundamental physical principles.To address this,we present a new approach using Physics-Integrated Neural Networks(PINNs),that combines data-driven learning with domain-specific physical constraints,embedding petrophysical relationships directly into the neural network architecture.This approach enforces that predictions adhere to physical laws.The methodology is applied to the Broom Creek Deep Saline aquifer,a CO_(2) sequestration site in the Williston Basin,to predict the volumes of key mineral constituents—quartz,dolomite,feldspar,anhydrite,illite—along with porosity.Compared to traditional artificial neural networks (ANN),the PINN approach demonstrates higher accuracy and better generalizability,significantly enhancing predictive performance on unseen well datasets.The average mean error across the three blind wells is 0.123 for ANN and 0.042 for PINN,highlighting the superior accuracy of the PINN approach.This method reduces uncertainties in reservoir characterization by improving the reliability of mineralogy and porosity predictions,providing a more robust tool for decision-making in various subsurface geoscience applications.展开更多
With the increasing integration of large-scale distributed energy resources into the grid,traditional distribution network optimization and dispatch methods struggle to address the challenges posed by both generation ...With the increasing integration of large-scale distributed energy resources into the grid,traditional distribution network optimization and dispatch methods struggle to address the challenges posed by both generation and load.Accounting for these issues,this paper proposes a multi-timescale coordinated optimization dispatch method for distribution networks.First,the probability box theory was employed to determine the uncertainty intervals of generation and load forecasts,based on which,the requirements for flexibility dispatch and capacity constraints of the grid were calculated and analyzed.Subsequently,a multi-timescale optimization framework was constructed,incorporating the generation and load forecast uncertainties.This framework included optimization models for dayahead scheduling,intra-day optimization,and real-time adjustments,aiming to meet flexibility needs across different timescales and improve the economic efficiency of the grid.Furthermore,an improved soft actor-critic algorithm was introduced to enhance the uncertainty exploration capability.Utilizing a centralized training and decentralized execution framework,a multi-agent SAC network model was developed to improve the decision-making efficiency of the agents.Finally,the effectiveness and superiority of the proposed method were validated using a modified IEEE-33 bus test system.展开更多
Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective meas...Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective measures,and designing civil defense engineering solutions.Current state-of-the-art methods face several issues:Experimental research is difficult and costly to implement,theoretical research is limited to simple geometries and lacks precision,and direct simulations require substantial computational resources.To address these challenges,this paper presents a data-driven method for predicting blast loads on building surfaces.This approach increases both the accuracy and computational efficiency of load predictions when the geometry of the building changes while the explosive yield remains constant,significantly improving its applicability in complex scenarios.This study introduces an innovative encoder-decoder graph neural network model named BlastGraphNet,which uses a message-passing mechanism to predict the overpressure and impulse load distributions on buildings with conventional and complex geometries during explosive events.The model also facilitates related downstream applications,such as damage mode identification and rapid assessment of virtual city explosions.The calculation results indicate that the prediction error of the model for conventional building tests is less than 2%,and its inference speed is 3-4 orders of magnitude faster than that of state-of-the-art numerical methods.In extreme test cases involving buildings with complex geometries and building clusters,the method achieved high accuracy and excellent generalizability.The strong adaptability and generalizability of BlastGraphNet confirm that this novel method enables precise real-time prediction of blast loads and provides a new paradigm for damage assessment in protective engineering.展开更多
文摘According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,given the incompleteness of the interactome and the limited knowledge of disease-associated genes,it is not obvious if the available data have sufficient coverage to map out modules associated with each disease.
文摘According to the disease module hypothesis,the cellular components associated with a disease segregate in the same neighborhood of the human interactome,the map of biologically relevant molecular interactions.Yet,given the incompleteness of the interactome and the limited knowledge of disease-associated genes,it is not obvious if the available data have sufficient coverage to map out modules associated with each disease.Here we derive mathematical conditions for the identifiability of disease modules and show that the network-based location of each disease module determines its pathobiological relationship to other diseases.For example,diseases with overlapping network modules show significant coexpression patterns,symptom similarity,and comorbidity,whereas diseases residing in separated network neighborhoods are phenotypically distinct.These tools represent an interactome-based platform to predict molecular commonalities between phenotypically related diseases,even if they do not share primary disease genes.
文摘Optimizing routing and resource allocation in decentralized unmanned aerial vehicle(UAV)networks remains challenging due to interference and rapidly changing topologies.The authors introduce a novel framework combining double deep Q-networks(DDQNs)and graph neural networks(GNNs)for joint routing and resource allocation.The framework uses GNNs to model the network topology and DDQNs to adaptively control routing and resource allocation,addressing interference and improving network performance.Simulation results show that the proposed approach outperforms traditional methods such as Closest-to-Destination(c2Dst),Max-SINR(mSINR),and Multi-Layer Perceptron(MLP)-based models,achieving approximately 23.5% improvement in throughput,50% increase in connection probability,and 17.6% reduction in number of hops,demonstrating its effectiveness in dynamic UAV networks.
基金supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(RS-2025-00559546)supported by the IITP(Institute of Information&Coummunications Technology Planning&Evaluation)-ITRC(Information Technology Research Center)grant funded by the Korea government(Ministry of Science and ICT)(IITP-2025-RS-2023-00259004).
文摘The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-generation(5G)networks transformed mobile broadband and machine-type communications at massive scales,their properties of scaling,interference management,and latency remain a limitation in dense high mobility settings.To overcome these limitations,artificial intelligence(AI)and unmanned aerial vehicles(UAVs)have emerged as potential solutions to develop versatile,dynamic,and energy-efficient communication systems.The study proposes an AI-based UAV architecture that utilizes cooperative reinforcement learning(CoRL)to manage an autonomous network.The UAVs collaborate by sharing local observations and real-time state exchanges to optimize user connectivity,movement directions,allocate power,and resource distribution.Unlike conventional centralized or autonomous methods,CoRL involves joint state sharing and conflict-sensitive reward shaping,which ensures fair coverage,less interference,and enhanced adaptability in a dynamic urban environment.Simulations conducted in smart city scenarios with 10 UAVs and 50 ground users demonstrate that the proposed CoRL-based UAV system increases user coverage by up to 10%,achieves convergence 40%faster,and reduces latency and energy consumption by 30%compared with centralized and decentralized baselines.Furthermore,the distributed nature of the algorithm ensures scalability and flexibility,making it well-suited for future large-scale 6G deployments.The results highlighted that AI-enabled UAV systems enhance connectivity,support ultra-reliable low-latency communications(URLLC),and improve 6G network efficiency.Future work will extend the framework with adaptive modulation,beamforming-aware positioning,and real-world testbed deployment.
文摘This paper investigates the traffic offloading optimization challenge in Space-Air-Ground Integrated Networks(SAGIN)through a novel Recursive Multi-Agent Proximal Policy Optimization(RMAPPO)algorithm.The exponential growth of mobile devices and data traffic has substantially increased network congestion,particularly in urban areas and regions with limited terrestrial infrastructure.Our approach jointly optimizes unmanned aerial vehicle(UAV)trajectories and satellite-assisted offloading strategies to simultaneously maximize data throughput,minimize energy consumption,and maintain equitable resource distribution.The proposed RMAPPO framework incorporates recurrent neural networks(RNNs)to model temporal dependencies in UAV mobility patterns and utilizes a decentralized multi-agent reinforcement learning architecture to reduce communication overhead while improving system robustness.The proposed RMAPPO algorithm was evaluated through simulation experiments,with the results indicating that it significantly enhances the cumulative traffic offloading rate of nodes and reduces the energy consumption of UAVs.
文摘Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01296).
文摘Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.
基金supported by the National Key Research and Development Program of China under Grant 2022YFB2901501in part by the Science and Technology Innovation leading Talents Subsidy Project of Central Plains under Grant 244200510038.
文摘The rapid growth of distributed data-centric applications and AI workloads increases demand for low-latency,high-throughput communication,necessitating frequent and flexible updates to network routing configurations.However,maintaining consistent forwarding states during these updates is challenging,particularly when rerouting multiple flows simultaneously.Existing approaches pay little attention to multi-flow update,where improper update sequences across data plane nodes may construct deadlock dependencies.Moreover,these methods typically involve excessive control-data plane interactions,incurring significant resource overhead and performance degradation.This paper presents P4LoF,an efficient loop-free update approach that enables the controller to reroute multiple flows through minimal interactions.P4LoF first utilizes a greedy-based algorithm to generate the shortest update dependency chain for the single-flow update.These chains are then dynamically merged into a dependency graph and resolved as a Shortest Common Super-sequence(SCS)problem to produce the update sequence of multi-flow update.To address deadlock dependencies in multi-flow updates,P4LoF builds a deadlock-fix forwarding model that leverages the flexible packet processing capabilities of the programmable data plane.Experimental results show that P4LoF reduces control-data plane interactions by at least 32.6%with modest overhead,while effectively guaranteeing loop-free consistency.
基金supported by the Science and Technology Project of Sichuan Electric Power Company“Power Supply Guarantee Strategy for Urban Distribution Networks Considering Coordination with Virtual Power Plant during Extreme Weather Event”(No.521920230003).
文摘Ensuring reliable power supply in urban distribution networks is a complex and critical task.To address the increased demand during extreme scenarios,this paper proposes an optimal dispatch strategy that considers the coordination with virtual power plants(VPPs).The proposed strategy improves systemflexibility and responsiveness by optimizing the power adjustment of flexible resources.In the proposed strategy,theGaussian Process Regression(GPR)is firstly employed to determine the adjustable range of aggregated power within the VPP,facilitating an assessment of its potential contribution to power supply support.Then,an optimal dispatch model based on a leader-follower game is developed to maximize the benefits of the VPP and flexible resources while guaranteeing the power balance at the same time.To solve the proposed optimal dispatch model efficiently,the constraints of the problem are reformulated and resolved using the Karush-Kuhn-Tucker(KKT)optimality conditions and linear programming duality theorem.The effectiveness of the strategy is illustrated through a detailed case study.
文摘Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the use of virtual reality(VR)technology.VR has been demonstrated to be an effective treatment for pain associated with medical procedures,as well as for chronic pain conditions for which no effective treatment has been established.The precise mechanism by which the diversion from reality facilitated by VR contributes to the diminution of pain and anxiety has yet to be elucidated.However,the provision of positive images through VR-based visual stimulation may enhance the functionality of brain networks.The salience network is diminished,while the default mode network is enhanced.Additionally,the medial prefrontal cortex may establish a stronger connection with the default mode network,which could result in a reduction of pain and anxiety.Further research into the potential of VR technology to alleviate pain could lead to a reduction in the number of individuals who overdose on painkillers and contribute to positive change in the medical field.
文摘Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.
基金National Natural Science Foundation of China(11971211,12171388).
文摘Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.
文摘Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.
文摘The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.
基金supported by the National Key Research and Development Program of China No.2023YFA1009500.
文摘With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.
基金supported by the Beijing Natural Science Foundation(Grant No.L223013)。
文摘For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models and physical models,each offering unique advantages but also facing limitations.Physics-informed neural networks(PINNs)provide a robust framework to integrate data-driven models with physical principles,ensuring consistency with underlying physics while enabling generalization across diverse operational conditions.This study introduces a PINN-based approach to reconstruct open circuit voltage(OCV)curves and estimate key ageing parameters at both the cell and electrode levels.These parameters include available capacity,electrode capacities,and lithium inventory capacity.The proposed method integrates OCV reconstruction models as functional components into convolutional neural networks(CNNs)and is validated using a public dataset.The results reveal that the estimated ageing parameters closely align with those obtained through offline OCV tests,with errors in reconstructed OCV curves remaining within 15 mV.This demonstrates the ability of the method to deliver fast and accurate degradation diagnostics at the electrode level,advancing the potential for precise and efficient battery health management.
基金the North Dakota Industrial Commission (NDIC) for their financial supportprovided by the University of North Dakota Computational Research Center。
文摘Accurate estimation of mineralogy from geophysical well logs is crucial for characterizing geological formations,particularly in hydrocarbon exploration,CO_(2) sequestration,and geothermal energy development.Current techniques,such as multimineral petrophysical analysis,offer details into mineralogical distribution.However,it is inherently time-intensive and demands substantial geological expertise for accurate model evaluation.Furthermore,traditional machine learning techniques often struggle to predict mineralogy accurately and sometimes produce estimations that violate fundamental physical principles.To address this,we present a new approach using Physics-Integrated Neural Networks(PINNs),that combines data-driven learning with domain-specific physical constraints,embedding petrophysical relationships directly into the neural network architecture.This approach enforces that predictions adhere to physical laws.The methodology is applied to the Broom Creek Deep Saline aquifer,a CO_(2) sequestration site in the Williston Basin,to predict the volumes of key mineral constituents—quartz,dolomite,feldspar,anhydrite,illite—along with porosity.Compared to traditional artificial neural networks (ANN),the PINN approach demonstrates higher accuracy and better generalizability,significantly enhancing predictive performance on unseen well datasets.The average mean error across the three blind wells is 0.123 for ANN and 0.042 for PINN,highlighting the superior accuracy of the PINN approach.This method reduces uncertainties in reservoir characterization by improving the reliability of mineralogy and porosity predictions,providing a more robust tool for decision-making in various subsurface geoscience applications.
基金funded by Jilin Province Science and Technology Development Plan Project,grant number 20220203163SF.
文摘With the increasing integration of large-scale distributed energy resources into the grid,traditional distribution network optimization and dispatch methods struggle to address the challenges posed by both generation and load.Accounting for these issues,this paper proposes a multi-timescale coordinated optimization dispatch method for distribution networks.First,the probability box theory was employed to determine the uncertainty intervals of generation and load forecasts,based on which,the requirements for flexibility dispatch and capacity constraints of the grid were calculated and analyzed.Subsequently,a multi-timescale optimization framework was constructed,incorporating the generation and load forecast uncertainties.This framework included optimization models for dayahead scheduling,intra-day optimization,and real-time adjustments,aiming to meet flexibility needs across different timescales and improve the economic efficiency of the grid.Furthermore,an improved soft actor-critic algorithm was introduced to enhance the uncertainty exploration capability.Utilizing a centralized training and decentralized execution framework,a multi-agent SAC network model was developed to improve the decision-making efficiency of the agents.Finally,the effectiveness and superiority of the proposed method were validated using a modified IEEE-33 bus test system.
基金supported by the National Natural Science Founion of China(U2241285).
文摘Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective measures,and designing civil defense engineering solutions.Current state-of-the-art methods face several issues:Experimental research is difficult and costly to implement,theoretical research is limited to simple geometries and lacks precision,and direct simulations require substantial computational resources.To address these challenges,this paper presents a data-driven method for predicting blast loads on building surfaces.This approach increases both the accuracy and computational efficiency of load predictions when the geometry of the building changes while the explosive yield remains constant,significantly improving its applicability in complex scenarios.This study introduces an innovative encoder-decoder graph neural network model named BlastGraphNet,which uses a message-passing mechanism to predict the overpressure and impulse load distributions on buildings with conventional and complex geometries during explosive events.The model also facilitates related downstream applications,such as damage mode identification and rapid assessment of virtual city explosions.The calculation results indicate that the prediction error of the model for conventional building tests is less than 2%,and its inference speed is 3-4 orders of magnitude faster than that of state-of-the-art numerical methods.In extreme test cases involving buildings with complex geometries and building clusters,the method achieved high accuracy and excellent generalizability.The strong adaptability and generalizability of BlastGraphNet confirm that this novel method enables precise real-time prediction of blast loads and provides a new paradigm for damage assessment in protective engineering.