Multilayer complex dynamical networks,characterized by the intricate topological connections and diverse hierarchical structures,present significant challenges in determining complete structural configurations due to ...Multilayer complex dynamical networks,characterized by the intricate topological connections and diverse hierarchical structures,present significant challenges in determining complete structural configurations due to the unique functional attributes and interaction patterns inherent to different layers.This paper addresses the critical question of whether structural information from a known layer can be used to reconstruct the unknown intralayer structure of a target layer within general weighted output-coupling multilayer networks.Building upon the generalized synchronization principle,we propose an innovative reconstruction method that incorporates two essential components in the design of structure observers,the cross-layer coupling modulator and the structural divergence term.A key advantage of the proposed reconstruction method lies in its flexibility to freely designate both the unknown target layer and the known reference layer from the general weighted output-coupling multilayer network.The reduced dependency on full-state observability enables more deployment in engineering applications with partial measurements.Numerical simulations are conducted to validate the effectiveness of the proposed structure reconstruction method.展开更多
Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce different...Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.展开更多
Nowadays,advances in communication technology and cloud computing have spawned a variety of smart mobile devices,which will generate a great amount of computing-intensive businesses,and require corresponding resources...Nowadays,advances in communication technology and cloud computing have spawned a variety of smart mobile devices,which will generate a great amount of computing-intensive businesses,and require corresponding resources of computation and communication.Multiaccess edge computing(MEC)can offload computing-intensive tasks to the nearby edge servers,which alleviates the pressure of devices.Ultra-dense network(UDN)can provide effective spectrum resources by deploying a large number of micro base stations.Furthermore,network slicing can support various applications in different communication scenarios.Therefore,this paper integrates the ultra-dense network slicing and the MEC technology,and introduces a hybrid computing offloading strategy in order to satisfy various quality of service(QoS)of edge devices.In order to dynamically allocate limited resources,the above problem is formulated as multiagent distributed deep reinforcement learning(DRL),which will achieve low overhead computation offloading strategy and real-time resource allocation decisions.In this context,federated learning is added to train DRL agents in a distributed manner,where each agent is dedicated to exploring actions composed of offloading decisions and allocating resources,so as to jointly optimize system delay and energy consumption.Simulation results show that the proposed learning algorithm has better performance compared with other strategies in literature.展开更多
To study the throughput scheduling problem under interference temperature in cognitive radio networks, an immune algorithm-based suboptimal method was proposed based on its NP-hard feature. The problem is modeled as a...To study the throughput scheduling problem under interference temperature in cognitive radio networks, an immune algorithm-based suboptimal method was proposed based on its NP-hard feature. The problem is modeled as a constrained optimization problem to maximize the total throughput of the secondary users( SUs). The mapping between the throughput scheduling problems and the immune algorithm is given. Suitable immune operators are designed such as binary antibody encoding, antibody initialization based on pre-knowledge, a proportional clone to its affinity and an adaptive mutation operator associated with the evolutionary generation. The simulation results showthat the proposed algorithm can obtain about 95% of the optimal throughput and operate with much lower liner computational complexity.展开更多
Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in us...Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.展开更多
Through empirical analysis of the global structure of the Worldwide Marine Transportation Network (WMTN), we find that the WMTN, a small-world network, exhibits an exponential-like degree distribution. We hereby inv...Through empirical analysis of the global structure of the Worldwide Marine Transportation Network (WMTN), we find that the WMTN, a small-world network, exhibits an exponential-like degree distribution. We hereby investigate the efficiency of the WMTN by employing a simple definition. Compared with many other transportation networks, the WMTN possesses relatively low efficiency. Furthermore, by exploring the relationship between the topological structure and the container throughput, we find that strong correlations exist among the container throughout the degree and the clustering coefficient. Also, considering the navigational process that a ship travels in a real shipping line, we obtain that the weight of a seaport is proportional to the total probability contributed by all the passing shipping lines.展开更多
In intelligent transportation system(ITS), the interworking of vehicular networks(VN) and cellular networks(CN) is proposed to provide high-data-rate services to vehicles. As the network access quality for CN and VN i...In intelligent transportation system(ITS), the interworking of vehicular networks(VN) and cellular networks(CN) is proposed to provide high-data-rate services to vehicles. As the network access quality for CN and VN is location related, mobile data offloading(MDO), which dynamically selects access networks for vehicles, should be considered with vehicle route planning to further improve the wireless data throughput of individual vehicles and to enhance the performance of the entire ITS. In this paper, we investigate joint MDO and route selection for an individual vehicle in a metropolitan scenario. We aim to improve the throughput of the target vehicle while guaranteeing its transportation efficiency requirements in terms of traveling time and distance. To achieve this objective, we first formulate the joint route and access network selection problem as a semi-Markov decision process(SMDP). Then we propose an optimal algorithm to calculate its optimal policy. To further reduce the computation complexity, we derive a suboptimal algorithm which reduces the action space. Simulation results demonstrate that the proposed optimal algorithm significantly outperforms the existing work in total throughput and the late arrival ratio.Moreover, the heuristic algorithm is able to substantially reduce the computation time with only slight performance degradation.展开更多
Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective meas...Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective measures,and designing civil defense engineering solutions.Current state-of-the-art methods face several issues:Experimental research is difficult and costly to implement,theoretical research is limited to simple geometries and lacks precision,and direct simulations require substantial computational resources.To address these challenges,this paper presents a data-driven method for predicting blast loads on building surfaces.This approach increases both the accuracy and computational efficiency of load predictions when the geometry of the building changes while the explosive yield remains constant,significantly improving its applicability in complex scenarios.This study introduces an innovative encoder-decoder graph neural network model named BlastGraphNet,which uses a message-passing mechanism to predict the overpressure and impulse load distributions on buildings with conventional and complex geometries during explosive events.The model also facilitates related downstream applications,such as damage mode identification and rapid assessment of virtual city explosions.The calculation results indicate that the prediction error of the model for conventional building tests is less than 2%,and its inference speed is 3-4 orders of magnitude faster than that of state-of-the-art numerical methods.In extreme test cases involving buildings with complex geometries and building clusters,the method achieved high accuracy and excellent generalizability.The strong adaptability and generalizability of BlastGraphNet confirm that this novel method enables precise real-time prediction of blast loads and provides a new paradigm for damage assessment in protective engineering.展开更多
With the rapid development of wearable electronic skin technology, flexible strain sensors have shown great application prospects in the fields of human motion and physiological signal detection, medical diagnostics, ...With the rapid development of wearable electronic skin technology, flexible strain sensors have shown great application prospects in the fields of human motion and physiological signal detection, medical diagnostics, and human-computer interaction owing to their outstanding sensing performance. This paper reports a strain sensor with synergistic conductive network, consisting of stable carbon nanotube dispersion (CNT) layer and brittle MXene layer by dip-coating and electrostatic self-assembly method, and breathable three-dimensional (3D) flexible substrate of thermoplastic polyurethane (TPU) fibrous membrane prepared through electrospinning technology. The MXene/CNT@PDA-TPU (MC@p-TPU) flexible strain sensor had excellent air permeability, wide operating range (0–450 %), high sensitivity (Gauge Factor, GFmax = 8089.7), ultra-low detection limit (0.05 %), rapid response and recovery times (40 ms/60 ms), and excellent cycle stability and durability (10,000 cycles). Given its superior strain sensing capabilities, this sensor can be applied in physiological signals detection, human motion pattern recognition, and driving exoskeleton robots. In addition, MC@p-TPU fibrous membrane also exhibited excellent photothermal conversion performance and can be used as a wearable photo-heater, which has far-reaching application potential in the photothermal therapy of human joint diseases.展开更多
Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act ...Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act on two targets exhibit strong therapeutic effects and advantages against mutations.In this study,a novel computational workflow was developed to design dual-target SARS-CoV-2 candidate inhibitors with the Envelope protein and Main protease selected as the two target proteins.The drug-like molecules of our self-constructed 3D scaffold database were used as high-throughput molecular docking probes for feature extraction of two target protein pockets.A multi-layer perceptron(MLP)was employed to embed the binding affinities into a latent space as conditional vectors to control conditional distribution.Utilizing a conditional generative neural network,cG-SchNet,with 3D Euclidean group(E3)symmetries,the conditional probability distributions of molecular 3D structures were acquired and a set of novel SARS-CoV-2 dual-target candidate inhibitors were generated.The 1D probability,2D joint probability,and 2D cumulative probability distribution results indicate that the generated sets are significantly enhanced compared to the training set in the high binding affinity area.Among the 201 generated molecules,42 molecules exhibited a sum binding affinity exceeding 17.0 kcal/mol while 9 of them having a sum binding affinity exceeding 19.0 kcal/mol,demonstrating structure diversity along with strong dual-target affinities,good absorption,distribution,metabolism,excretion,and toxicity(ADMET)properties,and ease of synthesis.Dual-target drugs are rare and difficult to find,and our“high-throughput docking-multi-conditional generation”workflow offers a wide range of options for designing or optimizing potent dual-target SARS-CoV-2 inhibitors.展开更多
In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this neces...In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this necessitates effective management of computation and wireless resources tailored to the requirements of various services.The heterogeneity of computation resources and interference among shared wireless resources pose significant coordination and management challenges.To solve these problems,this work provides an overview of multi-dimensional resource management in 6G SIG RAN,including computation and wireless resource.Firstly it provides with a review of current investigations on computation and wireless resource management and an analysis of existing deficiencies and challenges.Then focusing on the provided challenges,the work proposes an MEC-based computation resource management scheme and a mixed numerology-based wireless resource management scheme.Furthermore,it outlines promising future technologies,including joint model-driven and data-driven resource management technology,and blockchain-based resource management technology within the 6G SIG network.The work also highlights remaining challenges,such as reducing communication costs associated with unstable ground-to-satellite links and overcoming barriers posed by spectrum isolation.Overall,this comprehensive approach aims to pave the way for efficient and effective resource management in future 6G networks.展开更多
Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse ...Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.展开更多
In the hybrid LTE cellular network with D2D(Device-to-Device) communication, D2D communication technologies can improve the spectral efficiency significantly. However, the D2D users have to reutilize the spectrum whic...In the hybrid LTE cellular network with D2D(Device-to-Device) communication, D2D communication technologies can improve the spectral efficiency significantly. However, the D2D users have to reutilize the spectrum which is allocated to the cellular users. Therefore, the co-channel interference will be more complicated in the case of crosscell D2D communications. In this article, a novel spectrum allocation algorithm for inter-cell D2D communication considering the traffic load is proposed. The traffic load can be balanced by the proposed algorithm. Meanwhile D2D users can multiplex the spectrum allocated to a number of cellular users with a certain percentage to meet the requirements of Qo S of D2D communications and reduce the interference to cellular users. Finally, the simulation results demonstrate that the proposed algorithm can meet the needs of D2D users, balance the traffic load and improve the overall throughput of the system.展开更多
The exponentially increasing number of heterogeneous Internet of Things(Io T)devices(e.g.,Wi Fi and Zig Bee)crowed in the same ISM band(2.4 G)and recent advances in CrossTechnology Communications(CTC)motivate us to ex...The exponentially increasing number of heterogeneous Internet of Things(Io T)devices(e.g.,Wi Fi and Zig Bee)crowed in the same ISM band(2.4 G)and recent advances in CrossTechnology Communications(CTC)motivate us to explore more efficient data collection and maximize network throughput.CTC enables Wi Fi and Zig Bee devices to communicate directly without any hardware changes or gateway equipment,which sheds light on a more efficient data collection design.In this work,we propose a distributed algorithm,named Max Bee,to compute the maximum network throughput,which is formulated as a linear programming problem.Considering that the problem turns out to be non-convex and hard to solve exactly,we propose a distributed algorithm to solve nonlinear programming by using the dual decomposition method and gradient/subgradient algorithms.Through extensive simulations on different sets of deployed Zig Bee and Wi Fi devices,we observe that the proposed algorithm significantly increases the network throughput based on CTC for Sensor Networks.展开更多
This paper investigates the maximal achievable multi-rate throughput problem of a multicast ses-sion at the presence of network coding. Deviating from previous works which focus on single-rate network coding, our work...This paper investigates the maximal achievable multi-rate throughput problem of a multicast ses-sion at the presence of network coding. Deviating from previous works which focus on single-rate network coding, our work takes the heterogeneity of sinks into account and provides multiple data layers to address the problem. Firstly formulated is the maximal achievable throughput problem with the assumption that the data layers are independent and layer rates are static. It is proved that the problem in this case is, unfortunately, Non-deterministic Polynomial-time (NP)-hard. In addition, our formulation is extended to the problems with dependent layers and dynamic layers. Furthermore, the approximation algorithm which satisfies certain fair-ness is proposed.展开更多
Although cerebral neuroplasticity following amputation has been observed, little is understood about how network-level functional reorganization occurs in the brain following upper-limb amputation. The objective of th...Although cerebral neuroplasticity following amputation has been observed, little is understood about how network-level functional reorganization occurs in the brain following upper-limb amputation. The objective of this study was to analyze alterations in brain network functional connectivity(FC) in upper-limb amputees(ULAs). This observational study included 40 ULAs and 40 healthy control subjects;all participants underwent resting-state functional magnetic resonance imaging. Changes in intra-and inter-network FC in ULAs were quantified using independent component analysis and brain network FC analysis. We also analyzed the correlation between FC and clinical manifestations, such as pain. We identified 11 independent components using independent component analysis from all subjects. In ULAs, intra-network FC was decreased in the left precuneus(precuneus gyrus) within the dorsal attention network and left precentral(precentral gyrus) within the auditory network;but increased in the left Parietal_Inf(inferior parietal, but supramarginal and angular gyri) within the ventral sensorimotor network, right Cerebelum_Crus2(crus Ⅱ of cerebellum) and left Temporal_Mid(middle temporal gyrus) within the ventral attention network, and left Rolandic_Oper(rolandic operculum) within the auditory network. ULAs also showed decreased inter-network FCs between the dorsal sensorimotor network and ventral sensorimotor network, the dorsal sensorimotor network and right frontoparietal network, and the dorsal sensorimotor network and dorsal attention network. Correlation analyses revealed negative correlations between inter-network FC changes and residual limb pain and phantom limb pain scores, but positive correlations between inter-network FC changes and daily activity hours of stump limb. These results show that post-amputation plasticity in ULAs is not restricted to local remapping;rather, it also occurs at a network level across several cortical regions. This observation provides additional insights into the plasticity of brain networks after upper-limb amputation, and could contribute to identification of the mechanisms underlying post-amputation pain.展开更多
This paper investigates the maximum network through- put for resource-constrained space networks based on the delay and disruption-tolerant networking (DTN) architecture. Specifically, this paper proposes a methodol...This paper investigates the maximum network through- put for resource-constrained space networks based on the delay and disruption-tolerant networking (DTN) architecture. Specifically, this paper proposes a methodology for calculating the maximum network throughput of multiple transmission tasks under storage and delay constraints over a space network. A mixed-integer linear programming (MILP) is formulated to solve this problem. Simula- tions results show that the proposed methodology can successfully calculate the optimal throughput of a space network under storage and delay constraints, as well as a clear, monotonic relationship between end-to-end delay and the maximum network throughput under storage constraints. At the same time, the optimization re- sults shine light on the routing and transport protocol design in space communication, which can be used to obtain the optimal network throughput.展开更多
This paper investigates the power control and routing problem in the communication process of an energy harvesting(EH)multi-hop cognitive radio network(CRN).The secondary user(SU)nodes(i.e.,source node and relay nodes...This paper investigates the power control and routing problem in the communication process of an energy harvesting(EH)multi-hop cognitive radio network(CRN).The secondary user(SU)nodes(i.e.,source node and relay nodes)harvest energy from the environment and use the energy exclusively for transmitting data.The SU nodes(i.e.,relay nodes)on the path,store and forward the received data to the destination node.We consider a real world scenario where the EH-SU node has only local causal knowledge,i.e.,at any time,each EH-SU node only has knowledge of its own EH process,channel state and currently received data.In order to study the power and routing issues,an optimization problem that maximizes path throughput considering quality of service(QoS)and available energy constraints is proposed.To solve this optimization problem,we propose a hybrid game theory routing and power control algorithm(HGRPC).The EH-SU nodes on the same path cooperate with each other,but EH-SU nodes on the different paths compete with each other.By selecting the best next hop node,we find the best strategy that can maximize throughput.In addition,we have established four steps to achieve routing,i.e.,route discovery,route selection,route reply,and route maintenance.Compared with the direct transmission,HGRPC has advantages in longer distances and higher hop counts.The algorithm generates more energy,reduces energy consumption and increases predictable residual energy.In particular,the time complexity of HGRPC is analyzed and its convergence is proved.In simulation experiments,the performance(i.e.,throughput and bit error rate(BER))of HGRPC is evaluated.Finally,experimental results show that HGRPC has higher throughput,longer network life,less latency,and lower energy consumption.展开更多
Increasing the coverage and capacity of cellular networks by deploying additional base stations is one of the fundamental objectives of fifth-generation(5G)networks.However,it leads to performance degradation and huge...Increasing the coverage and capacity of cellular networks by deploying additional base stations is one of the fundamental objectives of fifth-generation(5G)networks.However,it leads to performance degradation and huge spectral consumption due to the massive densification of connected devices and simultaneous access demand.To meet these access conditions and improve Quality of Service,resource allocation(RA)should be carefully optimized.Traditionally,RA problems are nonconvex optimizations,which are performed using heuristic methods,such as genetic algorithm,particle swarm optimization,and simulated annealing.However,the application of these approaches remains computationally expensive and unattractive for dense cellular networks.Therefore,artificial intelligence algorithms are used to improve traditional RA mechanisms.Deep learning is a promising tool for addressing resource management problems in wireless communication.In this study,we investigate a double deep Q-network-based RA framework that maximizes energy efficiency(EE)and total network throughput in unmanned aerial vehicle(UAV)-assisted terrestrial networks.Specifically,the system is studied under the constraints of interference.However,the optimization problem is formulated as a mixed integer nonlinear program.Within this framework,we evaluated the effect of height and the number of UAVs on EE and throughput.Then,in accordance with the experimental results,we compare the proposed algorithm with several artificial intelligence methods.Simulation results indicate that the proposed approach can increase EE with a considerable throughput.展开更多
Due to highly dynamic topology caused by fast moving nodes the Vehicular ad-hoc network (VANET) results in the existence of transient communication links, which degrade the performance of developed protocols. Establis...Due to highly dynamic topology caused by fast moving nodes the Vehicular ad-hoc network (VANET) results in the existence of transient communication links, which degrade the performance of developed protocols. Established routes frequently become stale, and existing communication flows are interrupted, incurring delay and additional overhead. In this paper we propose a novel hybrid routing protocol, which is the combined between of the table-driven routing and on-demand routing in VANET. It makes fast convergence in routing process, minimal drop links, making more reliable links, and adaptive with changing of VANET topology. With neighbor table is updated instantaneously, and using strong neighbor for routing process makes route discovery process start whenever it received requirement, and through using route mechanism appropriately it reduces significantly route overhead at each node. The simulation results illustrate the outstanding properties of our proposed routing protocol.展开更多
基金Project supported by the National Natural Science Foun-dation of China(Grant No.62373197)the Natural Science Foundation of the Higher Education Institutions of Jiangsu Province,China(Grant No.23KJB120010)+1 种基金the Industry-University-Research Cooperation Project of Jiangsu Province,China(Grant No.BY20251038)the Cultivation and In-cubation Project of the College of Automation,Nanjing Uni-versity of Posts and Telecommunications.
文摘Multilayer complex dynamical networks,characterized by the intricate topological connections and diverse hierarchical structures,present significant challenges in determining complete structural configurations due to the unique functional attributes and interaction patterns inherent to different layers.This paper addresses the critical question of whether structural information from a known layer can be used to reconstruct the unknown intralayer structure of a target layer within general weighted output-coupling multilayer networks.Building upon the generalized synchronization principle,we propose an innovative reconstruction method that incorporates two essential components in the design of structure observers,the cross-layer coupling modulator and the structural divergence term.A key advantage of the proposed reconstruction method lies in its flexibility to freely designate both the unknown target layer and the known reference layer from the general weighted output-coupling multilayer network.The reduced dependency on full-state observability enables more deployment in engineering applications with partial measurements.Numerical simulations are conducted to validate the effectiveness of the proposed structure reconstruction method.
基金funded by National Research Council of Thailand(contract No.N42A671047).
文摘Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.
文摘Nowadays,advances in communication technology and cloud computing have spawned a variety of smart mobile devices,which will generate a great amount of computing-intensive businesses,and require corresponding resources of computation and communication.Multiaccess edge computing(MEC)can offload computing-intensive tasks to the nearby edge servers,which alleviates the pressure of devices.Ultra-dense network(UDN)can provide effective spectrum resources by deploying a large number of micro base stations.Furthermore,network slicing can support various applications in different communication scenarios.Therefore,this paper integrates the ultra-dense network slicing and the MEC technology,and introduces a hybrid computing offloading strategy in order to satisfy various quality of service(QoS)of edge devices.In order to dynamically allocate limited resources,the above problem is formulated as multiagent distributed deep reinforcement learning(DRL),which will achieve low overhead computation offloading strategy and real-time resource allocation decisions.In this context,federated learning is added to train DRL agents in a distributed manner,where each agent is dedicated to exploring actions composed of offloading decisions and allocating resources,so as to jointly optimize system delay and energy consumption.Simulation results show that the proposed learning algorithm has better performance compared with other strategies in literature.
基金The National Natural Science Foundation of China(No.U150461361202099+2 种基金61201175U1204618)China Postdoctoral Science Foundation(No.2013M541586)
文摘To study the throughput scheduling problem under interference temperature in cognitive radio networks, an immune algorithm-based suboptimal method was proposed based on its NP-hard feature. The problem is modeled as a constrained optimization problem to maximize the total throughput of the secondary users( SUs). The mapping between the throughput scheduling problems and the immune algorithm is given. Suitable immune operators are designed such as binary antibody encoding, antibody initialization based on pre-knowledge, a proportional clone to its affinity and an adaptive mutation operator associated with the evolutionary generation. The simulation results showthat the proposed algorithm can obtain about 95% of the optimal throughput and operate with much lower liner computational complexity.
文摘Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.
基金Supported by the National Natural Science Foundation of China under Grant Nos 70571027, 10647125, 10635020 and 70401020, and the Program of Introducing Talents of Discipline to Universities under Grant No B08033.
文摘Through empirical analysis of the global structure of the Worldwide Marine Transportation Network (WMTN), we find that the WMTN, a small-world network, exhibits an exponential-like degree distribution. We hereby investigate the efficiency of the WMTN by employing a simple definition. Compared with many other transportation networks, the WMTN possesses relatively low efficiency. Furthermore, by exploring the relationship between the topological structure and the container throughput, we find that strong correlations exist among the container throughout the degree and the clustering coefficient. Also, considering the navigational process that a ship travels in a real shipping line, we obtain that the weight of a seaport is proportional to the total probability contributed by all the passing shipping lines.
基金the National Natural Science Foundation of China under Grants 61631005 and U1801261the National Key R&D Program of China under Grant 2018YFB1801105+3 种基金the Central Universities under Grant ZYGX2019Z022the Key Areas of Research and Development Program of Guangdong Province, China, under Grant 2018B010114001the 111 Project under Grant B20064the China Postdoctoral Science Foundation under Grant No. 2018M631075
文摘In intelligent transportation system(ITS), the interworking of vehicular networks(VN) and cellular networks(CN) is proposed to provide high-data-rate services to vehicles. As the network access quality for CN and VN is location related, mobile data offloading(MDO), which dynamically selects access networks for vehicles, should be considered with vehicle route planning to further improve the wireless data throughput of individual vehicles and to enhance the performance of the entire ITS. In this paper, we investigate joint MDO and route selection for an individual vehicle in a metropolitan scenario. We aim to improve the throughput of the target vehicle while guaranteeing its transportation efficiency requirements in terms of traveling time and distance. To achieve this objective, we first formulate the joint route and access network selection problem as a semi-Markov decision process(SMDP). Then we propose an optimal algorithm to calculate its optimal policy. To further reduce the computation complexity, we derive a suboptimal algorithm which reduces the action space. Simulation results demonstrate that the proposed optimal algorithm significantly outperforms the existing work in total throughput and the late arrival ratio.Moreover, the heuristic algorithm is able to substantially reduce the computation time with only slight performance degradation.
基金supported by the National Natural Science Founion of China(U2241285).
文摘Accurate and efficient prediction of the distribution of surface loads on buildings subjected to explosive effects is crucial for rapidly calculating structural dynamic responses,establishing effective protective measures,and designing civil defense engineering solutions.Current state-of-the-art methods face several issues:Experimental research is difficult and costly to implement,theoretical research is limited to simple geometries and lacks precision,and direct simulations require substantial computational resources.To address these challenges,this paper presents a data-driven method for predicting blast loads on building surfaces.This approach increases both the accuracy and computational efficiency of load predictions when the geometry of the building changes while the explosive yield remains constant,significantly improving its applicability in complex scenarios.This study introduces an innovative encoder-decoder graph neural network model named BlastGraphNet,which uses a message-passing mechanism to predict the overpressure and impulse load distributions on buildings with conventional and complex geometries during explosive events.The model also facilitates related downstream applications,such as damage mode identification and rapid assessment of virtual city explosions.The calculation results indicate that the prediction error of the model for conventional building tests is less than 2%,and its inference speed is 3-4 orders of magnitude faster than that of state-of-the-art numerical methods.In extreme test cases involving buildings with complex geometries and building clusters,the method achieved high accuracy and excellent generalizability.The strong adaptability and generalizability of BlastGraphNet confirm that this novel method enables precise real-time prediction of blast loads and provides a new paradigm for damage assessment in protective engineering.
基金supported by the National Natural Science Foundation of China(Nos.52373093 and 12072325)the Outstanding Youth Fund of Henan Province(No.242300421062)+1 种基金National Key R&D Program of China(No.2019YFA0706802)the 111 project(No.D18023).
文摘With the rapid development of wearable electronic skin technology, flexible strain sensors have shown great application prospects in the fields of human motion and physiological signal detection, medical diagnostics, and human-computer interaction owing to their outstanding sensing performance. This paper reports a strain sensor with synergistic conductive network, consisting of stable carbon nanotube dispersion (CNT) layer and brittle MXene layer by dip-coating and electrostatic self-assembly method, and breathable three-dimensional (3D) flexible substrate of thermoplastic polyurethane (TPU) fibrous membrane prepared through electrospinning technology. The MXene/CNT@PDA-TPU (MC@p-TPU) flexible strain sensor had excellent air permeability, wide operating range (0–450 %), high sensitivity (Gauge Factor, GFmax = 8089.7), ultra-low detection limit (0.05 %), rapid response and recovery times (40 ms/60 ms), and excellent cycle stability and durability (10,000 cycles). Given its superior strain sensing capabilities, this sensor can be applied in physiological signals detection, human motion pattern recognition, and driving exoskeleton robots. In addition, MC@p-TPU fibrous membrane also exhibited excellent photothermal conversion performance and can be used as a wearable photo-heater, which has far-reaching application potential in the photothermal therapy of human joint diseases.
基金supported by Interdisciplinary Innova-tion Project of“Bioarchaeology Laboratory”of Jilin University,China,and“MedicineþX”Interdisciplinary Innovation Team of Norman Bethune Health Science Center of Jilin University,China(Grant No.:2022JBGS05).
文摘Severe acute respiratory syndrome coronavirus 2(SARS-CoV-2)mutations are influenced by random and uncontrollable factors,and the risk of the next widespread epidemic remains.Dual-target drugs that synergistically act on two targets exhibit strong therapeutic effects and advantages against mutations.In this study,a novel computational workflow was developed to design dual-target SARS-CoV-2 candidate inhibitors with the Envelope protein and Main protease selected as the two target proteins.The drug-like molecules of our self-constructed 3D scaffold database were used as high-throughput molecular docking probes for feature extraction of two target protein pockets.A multi-layer perceptron(MLP)was employed to embed the binding affinities into a latent space as conditional vectors to control conditional distribution.Utilizing a conditional generative neural network,cG-SchNet,with 3D Euclidean group(E3)symmetries,the conditional probability distributions of molecular 3D structures were acquired and a set of novel SARS-CoV-2 dual-target candidate inhibitors were generated.The 1D probability,2D joint probability,and 2D cumulative probability distribution results indicate that the generated sets are significantly enhanced compared to the training set in the high binding affinity area.Among the 201 generated molecules,42 molecules exhibited a sum binding affinity exceeding 17.0 kcal/mol while 9 of them having a sum binding affinity exceeding 19.0 kcal/mol,demonstrating structure diversity along with strong dual-target affinities,good absorption,distribution,metabolism,excretion,and toxicity(ADMET)properties,and ease of synthesis.Dual-target drugs are rare and difficult to find,and our“high-throughput docking-multi-conditional generation”workflow offers a wide range of options for designing or optimizing potent dual-target SARS-CoV-2 inhibitors.
基金supported by the National Key Research and Development Program of China(No.2021YFB2900504).
文摘In 6th Generation Mobile Networks(6G),the Space-Integrated-Ground(SIG)Radio Access Network(RAN)promises seamless coverage and exceptionally high Quality of Service(QoS)for diverse services.However,achieving this necessitates effective management of computation and wireless resources tailored to the requirements of various services.The heterogeneity of computation resources and interference among shared wireless resources pose significant coordination and management challenges.To solve these problems,this work provides an overview of multi-dimensional resource management in 6G SIG RAN,including computation and wireless resource.Firstly it provides with a review of current investigations on computation and wireless resource management and an analysis of existing deficiencies and challenges.Then focusing on the provided challenges,the work proposes an MEC-based computation resource management scheme and a mixed numerology-based wireless resource management scheme.Furthermore,it outlines promising future technologies,including joint model-driven and data-driven resource management technology,and blockchain-based resource management technology within the 6G SIG network.The work also highlights remaining challenges,such as reducing communication costs associated with unstable ground-to-satellite links and overcoming barriers posed by spectrum isolation.Overall,this comprehensive approach aims to pave the way for efficient and effective resource management in future 6G networks.
基金supported in part by the National Key R&D Program of China under Grant 2024YFE0200700in part by the National Natural Science Foundation of China under Grant 62201504.
文摘Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.
基金supported by the Open Research Fund of National Mobile Communications Research Laboratory,Southeast University(No.2015D07)
文摘In the hybrid LTE cellular network with D2D(Device-to-Device) communication, D2D communication technologies can improve the spectral efficiency significantly. However, the D2D users have to reutilize the spectrum which is allocated to the cellular users. Therefore, the co-channel interference will be more complicated in the case of crosscell D2D communications. In this article, a novel spectrum allocation algorithm for inter-cell D2D communication considering the traffic load is proposed. The traffic load can be balanced by the proposed algorithm. Meanwhile D2D users can multiplex the spectrum allocated to a number of cellular users with a certain percentage to meet the requirements of Qo S of D2D communications and reduce the interference to cellular users. Finally, the simulation results demonstrate that the proposed algorithm can meet the needs of D2D users, balance the traffic load and improve the overall throughput of the system.
基金supported by The Project funded by China Postdoctoral Science Foundation(Grant No.2018T110505,2017M611828)The Priority Academic Program Development(PAPD)of Jiangsu Higher Education Institutions。
文摘The exponentially increasing number of heterogeneous Internet of Things(Io T)devices(e.g.,Wi Fi and Zig Bee)crowed in the same ISM band(2.4 G)and recent advances in CrossTechnology Communications(CTC)motivate us to explore more efficient data collection and maximize network throughput.CTC enables Wi Fi and Zig Bee devices to communicate directly without any hardware changes or gateway equipment,which sheds light on a more efficient data collection design.In this work,we propose a distributed algorithm,named Max Bee,to compute the maximum network throughput,which is formulated as a linear programming problem.Considering that the problem turns out to be non-convex and hard to solve exactly,we propose a distributed algorithm to solve nonlinear programming by using the dual decomposition method and gradient/subgradient algorithms.Through extensive simulations on different sets of deployed Zig Bee and Wi Fi devices,we observe that the proposed algorithm significantly increases the network throughput based on CTC for Sensor Networks.
基金Supported by the National 863 High-tech Program of China (No.2003AA121560) and High-tech Project of Jiangsu Province (No.BG2003001).
文摘This paper investigates the maximal achievable multi-rate throughput problem of a multicast ses-sion at the presence of network coding. Deviating from previous works which focus on single-rate network coding, our work takes the heterogeneity of sinks into account and provides multiple data layers to address the problem. Firstly formulated is the maximal achievable throughput problem with the assumption that the data layers are independent and layer rates are static. It is proved that the problem in this case is, unfortunately, Non-deterministic Polynomial-time (NP)-hard. In addition, our formulation is extended to the problems with dependent layers and dynamic layers. Furthermore, the approximation algorithm which satisfies certain fair-ness is proposed.
基金supported by the National Natural Science Foundation of China, No.81974331(to XYZ)Shanghai Municipal Education Commission-Gaofeng Clinical Medicine Grant, No.20161429(to XYZ)
文摘Although cerebral neuroplasticity following amputation has been observed, little is understood about how network-level functional reorganization occurs in the brain following upper-limb amputation. The objective of this study was to analyze alterations in brain network functional connectivity(FC) in upper-limb amputees(ULAs). This observational study included 40 ULAs and 40 healthy control subjects;all participants underwent resting-state functional magnetic resonance imaging. Changes in intra-and inter-network FC in ULAs were quantified using independent component analysis and brain network FC analysis. We also analyzed the correlation between FC and clinical manifestations, such as pain. We identified 11 independent components using independent component analysis from all subjects. In ULAs, intra-network FC was decreased in the left precuneus(precuneus gyrus) within the dorsal attention network and left precentral(precentral gyrus) within the auditory network;but increased in the left Parietal_Inf(inferior parietal, but supramarginal and angular gyri) within the ventral sensorimotor network, right Cerebelum_Crus2(crus Ⅱ of cerebellum) and left Temporal_Mid(middle temporal gyrus) within the ventral attention network, and left Rolandic_Oper(rolandic operculum) within the auditory network. ULAs also showed decreased inter-network FCs between the dorsal sensorimotor network and ventral sensorimotor network, the dorsal sensorimotor network and right frontoparietal network, and the dorsal sensorimotor network and dorsal attention network. Correlation analyses revealed negative correlations between inter-network FC changes and residual limb pain and phantom limb pain scores, but positive correlations between inter-network FC changes and daily activity hours of stump limb. These results show that post-amputation plasticity in ULAs is not restricted to local remapping;rather, it also occurs at a network level across several cortical regions. This observation provides additional insights into the plasticity of brain networks after upper-limb amputation, and could contribute to identification of the mechanisms underlying post-amputation pain.
基金supported by the National Natural Sciences Foundation of China(6113200261321061+3 种基金6123101161201183)the National Basic Research Program of China(2014CB340206)the Tsinghua University Initiative Scientific Research Program(2011Z05117)
文摘This paper investigates the maximum network through- put for resource-constrained space networks based on the delay and disruption-tolerant networking (DTN) architecture. Specifically, this paper proposes a methodology for calculating the maximum network throughput of multiple transmission tasks under storage and delay constraints over a space network. A mixed-integer linear programming (MILP) is formulated to solve this problem. Simula- tions results show that the proposed methodology can successfully calculate the optimal throughput of a space network under storage and delay constraints, as well as a clear, monotonic relationship between end-to-end delay and the maximum network throughput under storage constraints. At the same time, the optimization re- sults shine light on the routing and transport protocol design in space communication, which can be used to obtain the optimal network throughput.
基金This work was partially supported by the National Natural Science Foundation of China(No.61771410,No.61876089)by the Postgraduate Innovation Fund Project by Southwest University of Science and Technology(No.19ycx0106)+2 种基金by the Artificial Intelligence Key Laboratory of Sichuan Province(No.2017RYY05,No.2018RYJ03)by the Zigong City Key Science and Technology Plan Project(2019YYJC16)by and by the Horizontal Project(No.HX2017134,No.HX2018264,Nos.E10203788,HX2019250).
文摘This paper investigates the power control and routing problem in the communication process of an energy harvesting(EH)multi-hop cognitive radio network(CRN).The secondary user(SU)nodes(i.e.,source node and relay nodes)harvest energy from the environment and use the energy exclusively for transmitting data.The SU nodes(i.e.,relay nodes)on the path,store and forward the received data to the destination node.We consider a real world scenario where the EH-SU node has only local causal knowledge,i.e.,at any time,each EH-SU node only has knowledge of its own EH process,channel state and currently received data.In order to study the power and routing issues,an optimization problem that maximizes path throughput considering quality of service(QoS)and available energy constraints is proposed.To solve this optimization problem,we propose a hybrid game theory routing and power control algorithm(HGRPC).The EH-SU nodes on the same path cooperate with each other,but EH-SU nodes on the different paths compete with each other.By selecting the best next hop node,we find the best strategy that can maximize throughput.In addition,we have established four steps to achieve routing,i.e.,route discovery,route selection,route reply,and route maintenance.Compared with the direct transmission,HGRPC has advantages in longer distances and higher hop counts.The algorithm generates more energy,reduces energy consumption and increases predictable residual energy.In particular,the time complexity of HGRPC is analyzed and its convergence is proved.In simulation experiments,the performance(i.e.,throughput and bit error rate(BER))of HGRPC is evaluated.Finally,experimental results show that HGRPC has higher throughput,longer network life,less latency,and lower energy consumption.
基金This work was supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2022R323)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia,and Taif University Researchers Supporting Project Number TURSP-2020/34),Taif,Saudi Arabia。
文摘Increasing the coverage and capacity of cellular networks by deploying additional base stations is one of the fundamental objectives of fifth-generation(5G)networks.However,it leads to performance degradation and huge spectral consumption due to the massive densification of connected devices and simultaneous access demand.To meet these access conditions and improve Quality of Service,resource allocation(RA)should be carefully optimized.Traditionally,RA problems are nonconvex optimizations,which are performed using heuristic methods,such as genetic algorithm,particle swarm optimization,and simulated annealing.However,the application of these approaches remains computationally expensive and unattractive for dense cellular networks.Therefore,artificial intelligence algorithms are used to improve traditional RA mechanisms.Deep learning is a promising tool for addressing resource management problems in wireless communication.In this study,we investigate a double deep Q-network-based RA framework that maximizes energy efficiency(EE)and total network throughput in unmanned aerial vehicle(UAV)-assisted terrestrial networks.Specifically,the system is studied under the constraints of interference.However,the optimization problem is formulated as a mixed integer nonlinear program.Within this framework,we evaluated the effect of height and the number of UAVs on EE and throughput.Then,in accordance with the experimental results,we compare the proposed algorithm with several artificial intelligence methods.Simulation results indicate that the proposed approach can increase EE with a considerable throughput.
基金Sponsored by the National Science and Technology Major Project(Grant No. 2010ZX03005-003)
文摘Due to highly dynamic topology caused by fast moving nodes the Vehicular ad-hoc network (VANET) results in the existence of transient communication links, which degrade the performance of developed protocols. Established routes frequently become stale, and existing communication flows are interrupted, incurring delay and additional overhead. In this paper we propose a novel hybrid routing protocol, which is the combined between of the table-driven routing and on-demand routing in VANET. It makes fast convergence in routing process, minimal drop links, making more reliable links, and adaptive with changing of VANET topology. With neighbor table is updated instantaneously, and using strong neighbor for routing process makes route discovery process start whenever it received requirement, and through using route mechanism appropriately it reduces significantly route overhead at each node. The simulation results illustrate the outstanding properties of our proposed routing protocol.