Internet of things and network densification bring significant challenges to uplink management.Only depending on optimization algorithm enhancements is not enough for uplink transmission.To control intercell interfere...Internet of things and network densification bring significant challenges to uplink management.Only depending on optimization algorithm enhancements is not enough for uplink transmission.To control intercell interference,Fractional Uplink Power Control(FUPC)should be optimized from network-wide perspective,which has to find a better traffic distribution model.Conventionally,traffic distribution is geographic-based,and ineffective due to tricky locating efforts.This paper proposes a novel uplink power management framework for Self-Organizing Networks(SON),which firstly builds up pathloss-based traffic distribution model and then makes the decision of FUPC based on the model.PathLoss-based Traffic Distribution(PLTD)aggregates traffic based on the propagation condition of traffic that is defined as the pathloss between the position generating the traffic and surrounding cells.Simulations show that the improvement in optimization efficiency of FUPC with PLTD can be up to 40%compared to conventional GeoGraphic-based Traffic Distribution(GGTD).展开更多
Decoupling of revenues with network traffic and extreme penetration of expenses in wireless network leads to the critical situation for wireless service providers (WSP), as more wireless network is complex due to its ...Decoupling of revenues with network traffic and extreme penetration of expenses in wireless network leads to the critical situation for wireless service providers (WSP), as more wireless network is complex due to its heterogeneity in the context of planning, software & hardware installation, radio parameters setting, drive testing, optimization, healing and maintenance. These operations are time-consuming, labor & budget-intensive and error-prone if activated manually. Hence new approaches have to be designed and applied to meet those demands in a cost-effective way, Self-organizing networks (SON), is a promising approach to handle manual tasks with autonomous manners. More specifically the self-directed functions (self-planning, self-deployment, self-configuration, self-optimization and self-healing) are aid to reduce capital expenditure (CAPEX), implementation expenditure (IMPEX) and operational expenditure (OPEX). In this study, first we investigate the aforementioned impact factors of cost combined with self-functions. Then, we analyze the relative cost benefits causing from deploying the SON functions, using the economical method to have more precise results concerning those potential benefits. At last, the result shows that there is a significant difference in expenses and revenues of WSP with and without SON after enabling self-functions in wireless network.展开更多
The information exchange among satellites is crucial for the implementation of cluster satellite cooperative missions.However,achieving fast perception,rapid networking,and highprecision time synchronization among nod...The information exchange among satellites is crucial for the implementation of cluster satellite cooperative missions.However,achieving fast perception,rapid networking,and highprecision time synchronization among nodes without the support of the Global Navigation Satellite System(GNSS)and other prior information remains a formidable challenge to real-time wireless networks design.Therefore,a self-organizing network methodology based on multi-agent negotiation is proposed,which autonomously determines the master node through collaborative negotiation and competitive elections.On this basis,a real-time network protocol design is carried out and a high-precision time synchronization method with motion compensation is proposed.Simulation results demonstrate that the proposed method enables rapid networking with the capability of selfdiscovery,self-organization,and self-healing.For a cluster of 8 satellites,the networking time and the reorganization time are less than 4 s.The time synchronization accuracy exceeds 10-10s with motion compensation,demonstrating excellent real-time performance and stability.The research presented in this paper provides a valuable reference for the design and application of spacebased self-organizing networks for satellite cluster.展开更多
Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the u...Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the use of virtual reality(VR)technology.VR has been demonstrated to be an effective treatment for pain associated with medical procedures,as well as for chronic pain conditions for which no effective treatment has been established.The precise mechanism by which the diversion from reality facilitated by VR contributes to the diminution of pain and anxiety has yet to be elucidated.However,the provision of positive images through VR-based visual stimulation may enhance the functionality of brain networks.The salience network is diminished,while the default mode network is enhanced.Additionally,the medial prefrontal cortex may establish a stronger connection with the default mode network,which could result in a reduction of pain and anxiety.Further research into the potential of VR technology to alleviate pain could lead to a reduction in the number of individuals who overdose on painkillers and contribute to positive change in the medical field.展开更多
Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently...Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.展开更多
Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in us...Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.展开更多
Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay di...Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.展开更多
Traveling salesman problem(TSP)is a classic non-deterministic polynomial-hard optimization prob-lem.Based on the characteristics of self-organizing mapping(SOM)network,this paper proposes an improved SOM network from ...Traveling salesman problem(TSP)is a classic non-deterministic polynomial-hard optimization prob-lem.Based on the characteristics of self-organizing mapping(SOM)network,this paper proposes an improved SOM network from the perspectives of network update strategy,initialization method,and parameter selection.This paper compares the performance of the proposed algorithms with the performance of existing SOM network algorithms on the TSP and compares them with several heuristic algorithms.Simulations show that compared with existing SOM networks,the improved SOM network proposed in this paper improves the convergence rate and algorithm accuracy.Compared with iterated local search and heuristic algorithms,the improved SOM net-work algorithms proposed in this paper have the advantage of fast calculation speed on medium-scale TSP.展开更多
For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models...For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models and physical models,each offering unique advantages but also facing limitations.Physics-informed neural networks(PINNs)provide a robust framework to integrate data-driven models with physical principles,ensuring consistency with underlying physics while enabling generalization across diverse operational conditions.This study introduces a PINN-based approach to reconstruct open circuit voltage(OCV)curves and estimate key ageing parameters at both the cell and electrode levels.These parameters include available capacity,electrode capacities,and lithium inventory capacity.The proposed method integrates OCV reconstruction models as functional components into convolutional neural networks(CNNs)and is validated using a public dataset.The results reveal that the estimated ageing parameters closely align with those obtained through offline OCV tests,with errors in reconstructed OCV curves remaining within 15 mV.This demonstrates the ability of the method to deliver fast and accurate degradation diagnostics at the electrode level,advancing the potential for precise and efficient battery health management.展开更多
With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based...With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.展开更多
The landscape of financial transactions has grown increasingly complex due to the expansion of global economic integration and advancements in information technology.This complexity poses greater challenges in detecti...The landscape of financial transactions has grown increasingly complex due to the expansion of global economic integration and advancements in information technology.This complexity poses greater challenges in detecting and managing financial fraud.This review explores the role of Graph Neural Networks(GNNs)in addressing these challenges by proposing a unified framework that categorizes existing GNN methodologies applied to financial fraud detection.Specifically,by examining a series of detailed research questions,this review delves into the suitability of GNNs for financial fraud detection,their deployment in real-world scenarios,and the design considerations that enhance their effectiveness.This review reveals that GNNs are exceptionally adept at capturing complex relational patterns and dynamics within financial networks,significantly outperforming traditional fraud detection methods.Unlike previous surveys that often overlook the specific potentials of GNNs or address them only superficially,our review provides a comprehensive,structured analysis,distinctly focusing on the multifaceted applications and deployments of GNNs in financial fraud detection.This review not only highlights the potential of GNNs to improve fraud detection mechanisms but also identifies current gaps and outlines future research directions to enhance their deployment in financial systems.Through a structured review of over 100 studies,this review paper contributes to the understanding of GNN applications in financial fraud detection,offering insights into their adaptability and potential integration strategies.展开更多
The increasing popularity of the Internet and the widespread use of information technology have led to a rise in the number and sophistication of network attacks and security threats.Intrusion detection systems are cr...The increasing popularity of the Internet and the widespread use of information technology have led to a rise in the number and sophistication of network attacks and security threats.Intrusion detection systems are crucial to network security,playing a pivotal role in safeguarding networks from potential threats.However,in the context of an evolving landscape of sophisticated and elusive attacks,existing intrusion detection methodologies often overlook critical aspects such as changes in network topology over time and interactions between hosts.To address these issues,this paper proposes a real-time network intrusion detection method based on graph neural networks.The proposedmethod leverages the advantages of graph neural networks and employs a straightforward graph construction method to represent network traffic as dynamic graph-structured data.Additionally,a graph convolution operation with a multi-head attention mechanism is utilized to enhance the model’s ability to capture the intricate relationships within the graph structure comprehensively.Furthermore,it uses an integrated graph neural network to address dynamic graphs’structural and topological changes at different time points and the challenges of edge embedding in intrusion detection data.The edge classification problem is effectively transformed into node classification by employing a line graph data representation,which facilitates fine-grained intrusion detection tasks on dynamic graph node feature representations.The efficacy of the proposed method is evaluated using two commonly used intrusion detection datasets,UNSW-NB15 and NF-ToN-IoT-v2,and results are compared with previous studies in this field.The experimental results demonstrate that our proposed method achieves 99.3%and 99.96%accuracy on the two datasets,respectively,and outperforms the benchmark model in several evaluation metrics.展开更多
The emergence of next generation networks(NextG),including 5G and beyond,is reshaping the technological landscape of cellular and mobile networks.These networks are sufficiently scaled to interconnect billions of user...The emergence of next generation networks(NextG),including 5G and beyond,is reshaping the technological landscape of cellular and mobile networks.These networks are sufficiently scaled to interconnect billions of users and devices.Researchers in academia and industry are focusing on technological advancements to achieve highspeed transmission,cell planning,and latency reduction to facilitate emerging applications such as virtual reality,the metaverse,smart cities,smart health,and autonomous vehicles.NextG continuously improves its network functionality to support these applications.Multiple input multiple output(MIMO)technology offers spectral efficiency,dependability,and overall performance in conjunctionwithNextG.This article proposes a secure channel estimation technique in MIMO topology using a norm-estimation model to provide comprehensive insights into protecting NextG network components against adversarial attacks.The technique aims to create long-lasting and secure NextG networks using this extended approach.The viability of MIMO applications and modern AI-driven methodologies to combat cybersecurity threats are explored in this research.Moreover,the proposed model demonstrates high performance in terms of reliability and accuracy,with a 20%reduction in the MalOut-RealOut-Diff metric compared to existing state-of-the-art techniques.展开更多
This study investigates photonuclear reaction(γ,n)cross-sections using Bayesian neural network(BNN)analysis.After determining the optimal network architecture,which features two hidden layers,each with 50 hidden node...This study investigates photonuclear reaction(γ,n)cross-sections using Bayesian neural network(BNN)analysis.After determining the optimal network architecture,which features two hidden layers,each with 50 hidden nodes,training was conducted for 30,000 iterations to ensure comprehensive data capture.By analyzing the distribution of absolute errors positively correlated with the cross-section for the isotope 159Tb,as well as the relative errors unrelated to the cross-section,we confirmed that the network effectively captured the data features without overfitting.Comparison with the TENDL-2021 Database demonstrated the BNN's reliability in fitting photonuclear cross-sections with lower average errors.The predictions for nuclei with single and double giant dipole resonance peak cross-sections,the accurate determination of the photoneutron reaction threshold in the low-energy region,and the precise description of trends in the high-energy cross-sections further demonstrate the network's generalization ability on the validation set.This can be attributed to the consistency of the training data.By using consistent training sets from different laboratories,Bayesian neural networks can predict nearby unknown cross-sections based on existing laboratory data,thereby estimating the potential differences between other laboratories'existing data and their own measurement results.Experimental measurements of photonuclear reactions on the newly constructed SLEGS beamline will contribute to clarifying the differences in cross-sections within the existing data.展开更多
Combined with elastic network model(ENM),the perturbation response scanning(PRS)has emerged as a robust technique for pinpointing allosteric interactions within proteins.Here,we proposed the PRS analysis of drug-targe...Combined with elastic network model(ENM),the perturbation response scanning(PRS)has emerged as a robust technique for pinpointing allosteric interactions within proteins.Here,we proposed the PRS analysis of drug-target networks(DTNs),which could provide a promising avenue in network medicine.We demonstrated the utility of the method by introducing a deep learning and network perturbation-based framework,for drug repurposing of multiple sclerosis(MS).First,the MS comorbidity network was constructed by performing a random walk with restart algorithm based on shared genes between MS and other diseases as seed nodes.Then,based on topological analysis and functional annotation,the neurotransmission module was identified as the“therapeutic module”of MS.Further,perturbation scores of drugs on the module were calculated by constructing the DTN and introducing the PRS analysis,giving a list of repurposable drugs for MS.Mechanism of action analysis both at pathway and structural levels screened dihydroergocristine as a candidate drug of MS by targeting a serotonin receptor of se-rotonin 2B receptor(HTR2B).Finally,we established a cuprizone-induced chronic mouse model to evaluate the alteration of HTR2B in mouse brain regions and observed that HTR2B was significantly reduced in the cuprizone-induced mouse cortex.These findings proved that the network perturbation modeling is a promising avenue for drug repurposing of MS.As a useful systematic method,our approach can also be used to discover the new molecular mechanism and provide effective candidate drugs for other complex diseases.展开更多
Spectrum-based fault localization (SBFL) generates a ranked list of suspicious elements by using the program execution spectrum, but the excessive number of elements ranked in parallel results in low localization accu...Spectrum-based fault localization (SBFL) generates a ranked list of suspicious elements by using the program execution spectrum, but the excessive number of elements ranked in parallel results in low localization accuracy. Most researchers consider intra-class dependencies to improve localization accuracy. However, some studies show that inter-class method call type faults account for more than 20%, which means such methods still have certain limitations. To solve the above problems, this paper proposes a two-phase software fault localization based on relational graph convolutional neural networks (Two-RGCNFL). Firstly, in Phase 1, the method call dependence graph (MCDG) of the program is constructed, the intra-class and inter-class dependencies in MCDG are extracted by using the relational graph convolutional neural network, and the classifier is used to identify the faulty methods. Then, the GraphSMOTE algorithm is improved to alleviate the impact of class imbalance on classification accuracy. Aiming at the problem of parallel ranking of element suspicious values in traditional SBFL technology, in Phase 2, Doc2Vec is used to learn static features, while spectrum information serves as dynamic features. A RankNet model based on siamese multi-layer perceptron is constructed to score and rank statements in the faulty method. This work conducts experiments on 5 real projects of Defects4J benchmark. Experimental results show that, compared with the traditional SBFL technique and two baseline methods, our approach improves the Top-1 accuracy by 262.86%, 29.59% and 53.01%, respectively, which verifies the effectiveness of Two-RGCNFL. Furthermore, this work verifies the importance of inter-class dependencies through ablation experiments.展开更多
In the context of the rapid iteration of information technology,the Internet of Things(IoT)has established itself as a pivotal hub connecting the digital world and the physical world.Wireless Sensor Networks(WSNs),dee...In the context of the rapid iteration of information technology,the Internet of Things(IoT)has established itself as a pivotal hub connecting the digital world and the physical world.Wireless Sensor Networks(WSNs),deeply embedded in the perception layer architecture of the IoT,play a crucial role as“tactile nerve endings.”A vast number of micro sensor nodes are widely distributed in monitoring areas according to preset deployment strategies,continuously and accurately perceiving and collecting real-time data on environmental parameters such as temperature,humidity,light intensity,air pressure,and pollutant concentration.These data are transmitted to the IoT cloud platform through stable and reliable communication links,forming a massive and detailed basic data resource pool.By using cutting-edge big data processing algorithms,machine learning models,and artificial intelligence analysis tools,in-depth mining and intelligent analysis of these multi-source heterogeneous data are conducted to generate high-value-added decision-making bases.This precisely empowers multiple fields,including agriculture,medical and health care,smart home,environmental science,and industrial manufacturing,driving intelligent transformation and catalyzing society to move towards a new stage of high-quality development.This paper comprehensively analyzes the technical cores of the IoT and WSNs,systematically sorts out the advanced key technologies of WSNs and the evolution of their strategic significance in the IoT system,deeply explores the innovative application scenarios and practical effects of the two in specific vertical fields,and looks forward to the technological evolution trends.It provides a detailed and highly practical guiding reference for researchers,technical engineers,and industrial decision-makers.展开更多
Software-defined networking(SDN)is an innovative paradigm that separates the control and data planes,introducing centralized network control.SDN is increasingly being adopted by Carrier Grade networks,offering enhance...Software-defined networking(SDN)is an innovative paradigm that separates the control and data planes,introducing centralized network control.SDN is increasingly being adopted by Carrier Grade networks,offering enhanced networkmanagement capabilities than those of traditional networks.However,because SDN is designed to ensure high-level service availability,it faces additional challenges.One of themost critical challenges is ensuring efficient detection and recovery from link failures in the data plane.Such failures can significantly impact network performance and lead to service outages,making resiliency a key concern for the effective adoption of SDN.Since the recovery process is intrinsically dependent on timely failure detection,this research surveys and analyzes the current literature on both failure detection and recovery approaches in SDN.The survey provides a critical comparison of existing failure detection techniques,highlighting their advantages and disadvantages.Additionally,it examines the current failure recovery methods,categorized as either restoration-based or protection-based,and offers a comprehensive comparison of their strengths and limitations.Lastly,future research challenges and directions are discussed to address the shortcomings of existing failure recovery methods.展开更多
Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act ...Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act on two targets exhibit strong therapeutic effects and advantages against mutations.In this study,a novel computational workflow was developed to design dual-target SARS-CoV-2 candidate inhibitors with the Envelope protein and Main protease selected as the two target proteins.The drug-like molecules of our self-constructed 3D scaffold database were used as high-throughput molecular docking probes for feature extraction of two target protein pockets.A multi-layer perceptron(MLP)was employed to embed the binding affinities into a latent space as conditional vectors to control conditional distribution.Utilizing a conditional generative neural network,cG-SchNet,with 3D Euclidean group(E3)symmetries,the conditional probability distributions of molecular 3D structures were acquired and a set of novel SARS-CoV-2 dual-target candidate inhibitors were generated.The 1D probability,2D joint probability,and 2D cumulative probability distribution results indicate that the generated sets are significantly enhanced compared to the training set in the high binding affinity area.Among the 201 generated molecules,42 molecules exhibited a sum binding affinity exceeding 17.0 kcal/mol while 9 of them having a sum binding affinity exceeding 19.0 kcal/mol,demonstrating structure diversity along with strong dual-target affinities,good absorption,distribution,metabolism,excretion,and toxicity(ADMET)properties,and ease of synthesis.Dual-target drugs are rare and difficult to find,and our“high-throughput docking-multi-conditional generation”workflow offers a wide range of options for designing or optimizing potent dual-target SARS-CoV-2 inhibitors.展开更多
文摘Internet of things and network densification bring significant challenges to uplink management.Only depending on optimization algorithm enhancements is not enough for uplink transmission.To control intercell interference,Fractional Uplink Power Control(FUPC)should be optimized from network-wide perspective,which has to find a better traffic distribution model.Conventionally,traffic distribution is geographic-based,and ineffective due to tricky locating efforts.This paper proposes a novel uplink power management framework for Self-Organizing Networks(SON),which firstly builds up pathloss-based traffic distribution model and then makes the decision of FUPC based on the model.PathLoss-based Traffic Distribution(PLTD)aggregates traffic based on the propagation condition of traffic that is defined as the pathloss between the position generating the traffic and surrounding cells.Simulations show that the improvement in optimization efficiency of FUPC with PLTD can be up to 40%compared to conventional GeoGraphic-based Traffic Distribution(GGTD).
文摘Decoupling of revenues with network traffic and extreme penetration of expenses in wireless network leads to the critical situation for wireless service providers (WSP), as more wireless network is complex due to its heterogeneity in the context of planning, software & hardware installation, radio parameters setting, drive testing, optimization, healing and maintenance. These operations are time-consuming, labor & budget-intensive and error-prone if activated manually. Hence new approaches have to be designed and applied to meet those demands in a cost-effective way, Self-organizing networks (SON), is a promising approach to handle manual tasks with autonomous manners. More specifically the self-directed functions (self-planning, self-deployment, self-configuration, self-optimization and self-healing) are aid to reduce capital expenditure (CAPEX), implementation expenditure (IMPEX) and operational expenditure (OPEX). In this study, first we investigate the aforementioned impact factors of cost combined with self-functions. Then, we analyze the relative cost benefits causing from deploying the SON functions, using the economical method to have more precise results concerning those potential benefits. At last, the result shows that there is a significant difference in expenses and revenues of WSP with and without SON after enabling self-functions in wireless network.
基金supported by the National Natural Science Foundation of China(No.62401597)the Natural Science Foundation of Hunan Province,China(No.2024JJ6469)the Scientific Research Project of National University of Defense Technology,China(No.ZK22-02)。
文摘The information exchange among satellites is crucial for the implementation of cluster satellite cooperative missions.However,achieving fast perception,rapid networking,and highprecision time synchronization among nodes without the support of the Global Navigation Satellite System(GNSS)and other prior information remains a formidable challenge to real-time wireless networks design.Therefore,a self-organizing network methodology based on multi-agent negotiation is proposed,which autonomously determines the master node through collaborative negotiation and competitive elections.On this basis,a real-time network protocol design is carried out and a high-precision time synchronization method with motion compensation is proposed.Simulation results demonstrate that the proposed method enables rapid networking with the capability of selfdiscovery,self-organization,and self-healing.For a cluster of 8 satellites,the networking time and the reorganization time are less than 4 s.The time synchronization accuracy exceeds 10-10s with motion compensation,demonstrating excellent real-time performance and stability.The research presented in this paper provides a valuable reference for the design and application of spacebased self-organizing networks for satellite cluster.
文摘Medical procedures are inherently invasive and carry the risk of inducing pain to the mind and body.Recently,efforts have been made to alleviate the discomfort associated with invasive medical procedures through the use of virtual reality(VR)technology.VR has been demonstrated to be an effective treatment for pain associated with medical procedures,as well as for chronic pain conditions for which no effective treatment has been established.The precise mechanism by which the diversion from reality facilitated by VR contributes to the diminution of pain and anxiety has yet to be elucidated.However,the provision of positive images through VR-based visual stimulation may enhance the functionality of brain networks.The salience network is diminished,while the default mode network is enhanced.Additionally,the medial prefrontal cortex may establish a stronger connection with the default mode network,which could result in a reduction of pain and anxiety.Further research into the potential of VR technology to alleviate pain could lead to a reduction in the number of individuals who overdose on painkillers and contribute to positive change in the medical field.
基金National Natural Science Foundation of China(11971211,12171388).
文摘Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.
文摘Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.
文摘Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data.
基金the National Natural Science Foundation of China (No.61627810)the National Science and Technology Major Program of China (No.2018YFB1305003)the National Defense Science and Technology Outstanding Youth Science Foundation (No.2017-JCJQ-ZQ-031)。
文摘Traveling salesman problem(TSP)is a classic non-deterministic polynomial-hard optimization prob-lem.Based on the characteristics of self-organizing mapping(SOM)network,this paper proposes an improved SOM network from the perspectives of network update strategy,initialization method,and parameter selection.This paper compares the performance of the proposed algorithms with the performance of existing SOM network algorithms on the TSP and compares them with several heuristic algorithms.Simulations show that compared with existing SOM networks,the improved SOM network proposed in this paper improves the convergence rate and algorithm accuracy.Compared with iterated local search and heuristic algorithms,the improved SOM net-work algorithms proposed in this paper have the advantage of fast calculation speed on medium-scale TSP.
基金supported by the Beijing Natural Science Foundation(Grant No.L223013)。
文摘For the diagnostics and health management of lithium-ion batteries,numerous models have been developed to understand their degradation characteristics.These models typically fall into two categories:data-driven models and physical models,each offering unique advantages but also facing limitations.Physics-informed neural networks(PINNs)provide a robust framework to integrate data-driven models with physical principles,ensuring consistency with underlying physics while enabling generalization across diverse operational conditions.This study introduces a PINN-based approach to reconstruct open circuit voltage(OCV)curves and estimate key ageing parameters at both the cell and electrode levels.These parameters include available capacity,electrode capacities,and lithium inventory capacity.The proposed method integrates OCV reconstruction models as functional components into convolutional neural networks(CNNs)and is validated using a public dataset.The results reveal that the estimated ageing parameters closely align with those obtained through offline OCV tests,with errors in reconstructed OCV curves remaining within 15 mV.This demonstrates the ability of the method to deliver fast and accurate degradation diagnostics at the electrode level,advancing the potential for precise and efficient battery health management.
基金supported by the National Key Research and Development Program of China No.2023YFA1009500.
文摘With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.
基金supported by the National Key R&D Program of China(No.2022YFB4501704)the National Natural Science Foundation of China(Grant No.62102287)the Shanghai Science and Technology Innovation Action Plan Project(Nos.22YS1400600 and 22511100700).
文摘The landscape of financial transactions has grown increasingly complex due to the expansion of global economic integration and advancements in information technology.This complexity poses greater challenges in detecting and managing financial fraud.This review explores the role of Graph Neural Networks(GNNs)in addressing these challenges by proposing a unified framework that categorizes existing GNN methodologies applied to financial fraud detection.Specifically,by examining a series of detailed research questions,this review delves into the suitability of GNNs for financial fraud detection,their deployment in real-world scenarios,and the design considerations that enhance their effectiveness.This review reveals that GNNs are exceptionally adept at capturing complex relational patterns and dynamics within financial networks,significantly outperforming traditional fraud detection methods.Unlike previous surveys that often overlook the specific potentials of GNNs or address them only superficially,our review provides a comprehensive,structured analysis,distinctly focusing on the multifaceted applications and deployments of GNNs in financial fraud detection.This review not only highlights the potential of GNNs to improve fraud detection mechanisms but also identifies current gaps and outlines future research directions to enhance their deployment in financial systems.Through a structured review of over 100 studies,this review paper contributes to the understanding of GNN applications in financial fraud detection,offering insights into their adaptability and potential integration strategies.
文摘The increasing popularity of the Internet and the widespread use of information technology have led to a rise in the number and sophistication of network attacks and security threats.Intrusion detection systems are crucial to network security,playing a pivotal role in safeguarding networks from potential threats.However,in the context of an evolving landscape of sophisticated and elusive attacks,existing intrusion detection methodologies often overlook critical aspects such as changes in network topology over time and interactions between hosts.To address these issues,this paper proposes a real-time network intrusion detection method based on graph neural networks.The proposedmethod leverages the advantages of graph neural networks and employs a straightforward graph construction method to represent network traffic as dynamic graph-structured data.Additionally,a graph convolution operation with a multi-head attention mechanism is utilized to enhance the model’s ability to capture the intricate relationships within the graph structure comprehensively.Furthermore,it uses an integrated graph neural network to address dynamic graphs’structural and topological changes at different time points and the challenges of edge embedding in intrusion detection data.The edge classification problem is effectively transformed into node classification by employing a line graph data representation,which facilitates fine-grained intrusion detection tasks on dynamic graph node feature representations.The efficacy of the proposed method is evaluated using two commonly used intrusion detection datasets,UNSW-NB15 and NF-ToN-IoT-v2,and results are compared with previous studies in this field.The experimental results demonstrate that our proposed method achieves 99.3%and 99.96%accuracy on the two datasets,respectively,and outperforms the benchmark model in several evaluation metrics.
基金funding from King Saud University through Researchers Supporting Project number(RSP2024R387),King Saud University,Riyadh,Saudi Arabia.
文摘The emergence of next generation networks(NextG),including 5G and beyond,is reshaping the technological landscape of cellular and mobile networks.These networks are sufficiently scaled to interconnect billions of users and devices.Researchers in academia and industry are focusing on technological advancements to achieve highspeed transmission,cell planning,and latency reduction to facilitate emerging applications such as virtual reality,the metaverse,smart cities,smart health,and autonomous vehicles.NextG continuously improves its network functionality to support these applications.Multiple input multiple output(MIMO)technology offers spectral efficiency,dependability,and overall performance in conjunctionwithNextG.This article proposes a secure channel estimation technique in MIMO topology using a norm-estimation model to provide comprehensive insights into protecting NextG network components against adversarial attacks.The technique aims to create long-lasting and secure NextG networks using this extended approach.The viability of MIMO applications and modern AI-driven methodologies to combat cybersecurity threats are explored in this research.Moreover,the proposed model demonstrates high performance in terms of reliability and accuracy,with a 20%reduction in the MalOut-RealOut-Diff metric compared to existing state-of-the-art techniques.
基金supported by National key research and development program(No.2022YFA1602404)the National Natural Science Foundation of China(Nos.12388102,12275338,12005280)the Key Laboratory of Nuclear Data foundation(No.JCKY2022201C152)。
文摘This study investigates photonuclear reaction(γ,n)cross-sections using Bayesian neural network(BNN)analysis.After determining the optimal network architecture,which features two hidden layers,each with 50 hidden nodes,training was conducted for 30,000 iterations to ensure comprehensive data capture.By analyzing the distribution of absolute errors positively correlated with the cross-section for the isotope 159Tb,as well as the relative errors unrelated to the cross-section,we confirmed that the network effectively captured the data features without overfitting.Comparison with the TENDL-2021 Database demonstrated the BNN's reliability in fitting photonuclear cross-sections with lower average errors.The predictions for nuclei with single and double giant dipole resonance peak cross-sections,the accurate determination of the photoneutron reaction threshold in the low-energy region,and the precise description of trends in the high-energy cross-sections further demonstrate the network's generalization ability on the validation set.This can be attributed to the consistency of the training data.By using consistent training sets from different laboratories,Bayesian neural networks can predict nearby unknown cross-sections based on existing laboratory data,thereby estimating the potential differences between other laboratories'existing data and their own measurement results.Experimental measurements of photonuclear reactions on the newly constructed SLEGS beamline will contribute to clarifying the differences in cross-sections within the existing data.
基金supported by the National Natural Science Foundation of China(Grant Nos.:32271292,31872723,32200778,and 22377089)the Jiangsu Students Innovation and Entrepre-neurship Training Program,China(Program No.:202210285081Z)+6 种基金the Project of MOE Key Laboratory of Geriatric Diseases and Immunology,China(Project No.:JYN202404)Proj-ect Funded by the Priority Academic Program Development(PAPD)of Jiangsu Higher Education Institutions,Natural Science Foundation of Jiangsu Province,China(Project No.:BK20220494)Suzhou Medical and Health Technology Innovation Project,China(Grant No.:SKY2022107)the Clinical Research Center of Neuro-logical Disease in The Second Affiliated Hospital of Soochow University,China(Grant No.:ND2022A04)State Key Laboratory of Drug Research(Grant No.:SKLDR-2023-KF-05)Jiangsu Shuang-chuang Program for Doctor,Young Science Talents Promotion Project of Jiangsu Science and Technology Association(Program No.:TJ-2023-019)Young Science Talents Promotion Project of Suzhou Science and Technology Association,Suzhou International Joint Laboratory for Diagnosis and Treatment of Brain Diseases,and startup funding(Grant Nos.:NH21500221,NH21500122,and NH21500123)to Qifei Cong.
文摘Combined with elastic network model(ENM),the perturbation response scanning(PRS)has emerged as a robust technique for pinpointing allosteric interactions within proteins.Here,we proposed the PRS analysis of drug-target networks(DTNs),which could provide a promising avenue in network medicine.We demonstrated the utility of the method by introducing a deep learning and network perturbation-based framework,for drug repurposing of multiple sclerosis(MS).First,the MS comorbidity network was constructed by performing a random walk with restart algorithm based on shared genes between MS and other diseases as seed nodes.Then,based on topological analysis and functional annotation,the neurotransmission module was identified as the“therapeutic module”of MS.Further,perturbation scores of drugs on the module were calculated by constructing the DTN and introducing the PRS analysis,giving a list of repurposable drugs for MS.Mechanism of action analysis both at pathway and structural levels screened dihydroergocristine as a candidate drug of MS by targeting a serotonin receptor of se-rotonin 2B receptor(HTR2B).Finally,we established a cuprizone-induced chronic mouse model to evaluate the alteration of HTR2B in mouse brain regions and observed that HTR2B was significantly reduced in the cuprizone-induced mouse cortex.These findings proved that the network perturbation modeling is a promising avenue for drug repurposing of MS.As a useful systematic method,our approach can also be used to discover the new molecular mechanism and provide effective candidate drugs for other complex diseases.
基金funded by the Youth Fund of the National Natural Science Foundation of China(Grant No.42261070).
文摘Spectrum-based fault localization (SBFL) generates a ranked list of suspicious elements by using the program execution spectrum, but the excessive number of elements ranked in parallel results in low localization accuracy. Most researchers consider intra-class dependencies to improve localization accuracy. However, some studies show that inter-class method call type faults account for more than 20%, which means such methods still have certain limitations. To solve the above problems, this paper proposes a two-phase software fault localization based on relational graph convolutional neural networks (Two-RGCNFL). Firstly, in Phase 1, the method call dependence graph (MCDG) of the program is constructed, the intra-class and inter-class dependencies in MCDG are extracted by using the relational graph convolutional neural network, and the classifier is used to identify the faulty methods. Then, the GraphSMOTE algorithm is improved to alleviate the impact of class imbalance on classification accuracy. Aiming at the problem of parallel ranking of element suspicious values in traditional SBFL technology, in Phase 2, Doc2Vec is used to learn static features, while spectrum information serves as dynamic features. A RankNet model based on siamese multi-layer perceptron is constructed to score and rank statements in the faulty method. This work conducts experiments on 5 real projects of Defects4J benchmark. Experimental results show that, compared with the traditional SBFL technique and two baseline methods, our approach improves the Top-1 accuracy by 262.86%, 29.59% and 53.01%, respectively, which verifies the effectiveness of Two-RGCNFL. Furthermore, this work verifies the importance of inter-class dependencies through ablation experiments.
文摘In the context of the rapid iteration of information technology,the Internet of Things(IoT)has established itself as a pivotal hub connecting the digital world and the physical world.Wireless Sensor Networks(WSNs),deeply embedded in the perception layer architecture of the IoT,play a crucial role as“tactile nerve endings.”A vast number of micro sensor nodes are widely distributed in monitoring areas according to preset deployment strategies,continuously and accurately perceiving and collecting real-time data on environmental parameters such as temperature,humidity,light intensity,air pressure,and pollutant concentration.These data are transmitted to the IoT cloud platform through stable and reliable communication links,forming a massive and detailed basic data resource pool.By using cutting-edge big data processing algorithms,machine learning models,and artificial intelligence analysis tools,in-depth mining and intelligent analysis of these multi-source heterogeneous data are conducted to generate high-value-added decision-making bases.This precisely empowers multiple fields,including agriculture,medical and health care,smart home,environmental science,and industrial manufacturing,driving intelligent transformation and catalyzing society to move towards a new stage of high-quality development.This paper comprehensively analyzes the technical cores of the IoT and WSNs,systematically sorts out the advanced key technologies of WSNs and the evolution of their strategic significance in the IoT system,deeply explores the innovative application scenarios and practical effects of the two in specific vertical fields,and looks forward to the technological evolution trends.It provides a detailed and highly practical guiding reference for researchers,technical engineers,and industrial decision-makers.
文摘Software-defined networking(SDN)is an innovative paradigm that separates the control and data planes,introducing centralized network control.SDN is increasingly being adopted by Carrier Grade networks,offering enhanced networkmanagement capabilities than those of traditional networks.However,because SDN is designed to ensure high-level service availability,it faces additional challenges.One of themost critical challenges is ensuring efficient detection and recovery from link failures in the data plane.Such failures can significantly impact network performance and lead to service outages,making resiliency a key concern for the effective adoption of SDN.Since the recovery process is intrinsically dependent on timely failure detection,this research surveys and analyzes the current literature on both failure detection and recovery approaches in SDN.The survey provides a critical comparison of existing failure detection techniques,highlighting their advantages and disadvantages.Additionally,it examines the current failure recovery methods,categorized as either restoration-based or protection-based,and offers a comprehensive comparison of their strengths and limitations.Lastly,future research challenges and directions are discussed to address the shortcomings of existing failure recovery methods.
基金supported by Interdisciplinary Innova-tion Project of“Bioarchaeology Laboratory”of Jilin University,China,and“MedicineþX”Interdisciplinary Innovation Team of Norman Bethune Health Science Center of Jilin University,China(Grant No.:2022JBGS05).
文摘Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act on two targets exhibit strong therapeutic effects and advantages against mutations.In this study,a novel computational workflow was developed to design dual-target SARS-CoV-2 candidate inhibitors with the Envelope protein and Main protease selected as the two target proteins.The drug-like molecules of our self-constructed 3D scaffold database were used as high-throughput molecular docking probes for feature extraction of two target protein pockets.A multi-layer perceptron(MLP)was employed to embed the binding affinities into a latent space as conditional vectors to control conditional distribution.Utilizing a conditional generative neural network,cG-SchNet,with 3D Euclidean group(E3)symmetries,the conditional probability distributions of molecular 3D structures were acquired and a set of novel SARS-CoV-2 dual-target candidate inhibitors were generated.The 1D probability,2D joint probability,and 2D cumulative probability distribution results indicate that the generated sets are significantly enhanced compared to the training set in the high binding affinity area.Among the 201 generated molecules,42 molecules exhibited a sum binding affinity exceeding 17.0 kcal/mol while 9 of them having a sum binding affinity exceeding 19.0 kcal/mol,demonstrating structure diversity along with strong dual-target affinities,good absorption,distribution,metabolism,excretion,and toxicity(ADMET)properties,and ease of synthesis.Dual-target drugs are rare and difficult to find,and our“high-throughput docking-multi-conditional generation”workflow offers a wide range of options for designing or optimizing potent dual-target SARS-CoV-2 inhibitors.