Active distribution network(ADN)planning is crucial for achieving a cost-effective transition to modern power systems,yet it poses significant challenges as the system scale increases.The advent of quantum computing o...Active distribution network(ADN)planning is crucial for achieving a cost-effective transition to modern power systems,yet it poses significant challenges as the system scale increases.The advent of quantum computing offers a transformative approach to solve ADN planning.To fully leverage the potential of quantum computing,this paper proposes a photonic quantum acceleration algorithm.First,a quantum-accelerated framework for ADN planning is proposed on the basis of coherent photonic quantum computers.The ADN planning model is then formulated and decomposed into discrete master problems and continuous subproblems to facilitate the quantum optimization process.The photonic quantum-embedded adaptive alternating direction method of multipliers(PQA-ADMM)algorithm is subsequently proposed to equivalently map the discrete master problem onto a quantum-interpretable model,enabling its deployment on a photonic quantum computer.Finally,a comparative analysis with various solvers,including Gurobi,demonstrates that the proposed PQA-ADMM algorithm achieves significant speedup on the modified IEEE 33-node and IEEE 123-node systems,highlighting its effectiveness.展开更多
The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreser...The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreserving computation.Classical MPC relies on cryptographic techniques such as homomorphic encryption,secret sharing,and oblivious transfer,which may become vulnerable in the post-quantum era due to the computational power of quantum adversaries.This study presents a review of 140 peer-reviewed articles published between 2000 and 2025 that used different databases like MDPI,IEEE Explore,Springer,and Elsevier,examining the applications,types,and security issues with the solution of Quantum computing in different fields.This review explores the impact of quantum computing on MPC security,assesses emerging quantum-resistant MPC protocols,and examines hybrid classicalquantum approaches aimed at mitigating quantum threats.We analyze the role of Quantum Key Distribution(QKD),post-quantum cryptography(PQC),and quantum homomorphic encryption in securing multiparty computations.Additionally,we discuss the challenges of scalability,computational efficiency,and practical deployment of quantumsecure MPC frameworks in real-world applications such as privacy-preserving AI,secure blockchain transactions,and confidential data analysis.This review provides insights into the future research directions and open challenges in ensuring secure,scalable,and quantum-resistant multiparty computation.展开更多
The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous c...The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.展开更多
Organic electrochemical transistor(OECT)devices demonstrate great promising potential for reservoir computing(RC)systems,but their lack of tunable dynamic characteristics limits their application in multi-temporal sca...Organic electrochemical transistor(OECT)devices demonstrate great promising potential for reservoir computing(RC)systems,but their lack of tunable dynamic characteristics limits their application in multi-temporal scale tasks.In this study,we report an OECT-based neuromorphic device with tunable relaxation time(τ)by introducing an additional vertical back-gate electrode into a planar structure.The dual-gate design enablesτreconfiguration from 93 to 541 ms.The tunable relaxation behaviors can be attributed to the combined effects of planar-gate induced electrochemical doping and back-gateinduced electrostatic coupling,as verified by electrochemical impedance spectroscopy analysis.Furthermore,we used theτ-tunable OECT devices as physical reservoirs in the RC system for intelligent driving trajectory prediction,achieving a significant improvement in prediction accuracy from below 69%to 99%.The results demonstrate that theτ-tunable OECT shows a promising candidate for multi-temporal scale neuromorphic computing applications.展开更多
The deceleration of Moore's law and the energy–latency drawbacks of the von Neumann bottleneck have heightened the pursuit for beyond-CMOS designs that integrate memory and compute.Self-rectifying memristors(SRMs...The deceleration of Moore's law and the energy–latency drawbacks of the von Neumann bottleneck have heightened the pursuit for beyond-CMOS designs that integrate memory and compute.Self-rectifying memristors(SRMs)have emerged as promising building blocks for high-performance,low-power systems by combining resistive switching with intrinsic diode-like behavior.Their unidirectional conduction inhibits sneak-path currents in crossbar arrays devoid of external selectors,while nonlinear I–V characteristics,adjustable conductance states,low operating voltages,and rapid switching facilitate efficient vector–matrix operations,neuromorphic plasticity,and hardware security primitives.This review synthesizes the working mechanisms of SRMs,surveys material,and structural strategies and compares device metrics relevant to array-scale deployment(rectification ratio,nonlinearity,endurance,retention,variability,and operating voltage).We assess SRM-enabled in-memory computing and neuromorphic applications,as well as security functions such as physical unclonable functions and reconfigurable cryptographic primitives.Integration pathways toward CMOS compatibility are analyzed,including back-end-of-line thermal budgets,uniformity,write disturb mitigation,and reliability.Finally,we outline key challenges and opportunities:materials/architecture co-design,precision analog training,stochasticity control/exploitation,3D stacking,and standardized benchmarking that can accelerate large-scale SRM adoption.Through the use of specialized materials and structural optimization,SRMs are set to provide selector-free,densely integrated,and energy-efficient hardware for future information processing.展开更多
Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrain...Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.展开更多
The integration of large-scale foundation models(e.g.,GPT series and AlphaFold)into oncology is fundamentally transforming both research methodologies and clinical practices,driven by unprecedented advancements in com...The integration of large-scale foundation models(e.g.,GPT series and AlphaFold)into oncology is fundamentally transforming both research methodologies and clinical practices,driven by unprecedented advancements in computational power.This review synthesizes recent progress in the application of large language models to core oncological tasks,including medical imaging analysis,genomic interpretation,and personalized treatment planning.Underpinned by advanced computational infrastructures,such as graphics processing unit/tensor processing unit clusters,heterogeneous computing,and cloud platforms,these models enable superior representation learning and generalization across multimodal data sources.This review examines how these infrastructures overcome key bottlenecks in intelligent oncology through scalable optimization strategies,including mixed-precision training,memory optimization,and heterogeneous computing.Alongside these technical advancements,the review explores pressing challenges,such as data heterogeneity,limited model interpretability,regulatory uncertainties,and the environmental impact of artificial intelligence(AI)systems.Special emphasis is placed on emerging solutions,encompassing green AI and edge computing,which offer promising approaches for low-resource deployment scenarios.Additionally,the review highlights the critical role of interdisciplinary collaboration among oncology,computer science,ethics,and policy to ensure that AI systems are not only powerful but also transparent,safe,and clinically relevant.Finally,the review outlines potential avenues for future research aimed at developing robust,scalable,and human-centered frameworks for intelligent oncology.展开更多
In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network e...In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network environments where rapid response is mandatory for safety-critical applications.In scenarios where edge servers are sparsely deployed,the lack of coordination and information sharing often leads to load imbalance,thereby increasing system latency.Furthermore,in regions without edge server coverage,tasks must be processed locally,which further exacerbates latency issues.To address these challenges,we propose a novel and efficient Deep Reinforcement Learning(DRL)-based approach aimed at minimizing average task latency.The proposed method incorporates three offloading strategies:local computation,direct offloading to the edge server in local region,and device-to-device(D2D)-assisted offloading to edge servers in other regions.We formulate the task offloading process as a complex latency minimization optimization problem.To solve it,we propose an advanced algorithm based on the Dueling Double Deep Q-Network(D3QN)architecture and incorporating the Prioritized Experience Replay(PER)mechanism.Experimental results demonstrate that,compared with existing offloading algorithms,the proposed method significantly reduces average task latency,enhances user experience,and offers an effective strategy for latency optimization in future edge computing systems under dynamic workloads.展开更多
Rotational computed laminography(CL)has broad application potential in three-dimensional imaging of plate-like objects because it only requires X-rays to pass through the tested object in the thickness direction durin...Rotational computed laminography(CL)has broad application potential in three-dimensional imaging of plate-like objects because it only requires X-rays to pass through the tested object in the thickness direction during the imaging process.In this study,a rectangular cross-section field-of-view rotational CL(RC-CL)is proposed for circuit board imaging.Compared to other rotational CL systems,the field of view is the largest and most suitable for rectangular circuit boards.Meanwhile,as the imaging geometry of RC-CL is significantly different from that of cone-beam CT,the Feldkamp-Davis-Kress(FDK)reconstruction algorithm cannot be used directly.However,transferring the projection data to fit into the CBCT geometry using two-dimensional interpolation introduces interpolation errors.Therefore,an FDK-type analytical reconstruction algorithm applicable to RC-CL was developed.The effectiveness of the method was validated through numerical experiments,and the influence of the tilt angle on the reconstruction results was analyzed.Finally,the RC-CL technique was applied to real defect detection research on circuit boards.展开更多
In recent years,fog computing has become an important environment for dealing with the Internet of Things.Fog computing was developed to handle large-scale big data by scheduling tasks via cloud computing.Task schedul...In recent years,fog computing has become an important environment for dealing with the Internet of Things.Fog computing was developed to handle large-scale big data by scheduling tasks via cloud computing.Task scheduling is crucial for efficiently handling IoT user requests,thereby improving system performance,cost,and energy consumption across nodes in cloud computing.With the large amount of data and user requests,achieving the optimal solution to the task scheduling problem is challenging,particularly in terms of cost and energy efficiency.In this paper,we develop novel strategies to save energy consumption across nodes in fog computing when users execute tasks through the least-cost paths.Task scheduling is developed using modified artificial ecosystem optimization(AEO),combined with negative swarm operators,Salp Swarm Algorithm(SSA),in order to competitively optimize their capabilities during the exploitation phase of the optimal search process.In addition,the proposed strategy,Enhancement Artificial Ecosystem Optimization Salp Swarm Algorithm(EAEOSSA),attempts to find the most suitable solution.The optimization that combines cost and energy for multi-objective task scheduling optimization problems.The backpack problem is also added to improve both cost and energy in the iFogSim implementation as well.A comparison was made between the proposed strategy and other strategies in terms of time,cost,energy,and productivity.Experimental results showed that the proposed strategy improved energy consumption,cost,and time over other algorithms.Simulation results demonstrate that the proposed algorithm increases the average cost,average energy consumption,and mean service time in most scenarios,with average reductions of up to 21.15%in cost and 25.8%in energy consumption.展开更多
With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on comput...With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on computing latency and resource supply efficiency.Multi-access edge computing technology deploys cloud computing capabilities at the network edge;constructs distributed computing nodes and multi-access systems and offers infrastructure support for services with low latency and high reliability.Existing research relies on a strong assumption that the environmental state is fully observable and fails to thoroughly consider the continuous time-varying features of edge server load fluctuations,leading to insufficient adaptability of the model in a heterogeneous dynamic environment.Thus,this paper establishes a framework for end-edge collaborative task offloading based on a partially observable Markov decision-making process(POMDP)and proposes a method for end-edge collaborative task offloading in heterogeneous scenarios.It achieves time-series modeling of the historical load characteristics of edge servers and endows the agent with the ability to be aware of the load in dynamic environmental states.Moreover,by dynamically assessing the exploration value of historical trajectories in the central trajectory pool and adjusting the sample weight distribution,directional exploration and strategy optimization of high-value trajectories are realized.Experimental results indicate that the proposed method exhibits distinct advantages compared with existing methods in terms of average delay and task failure rate and also verifies the method’s robustness in a dynamic environment.展开更多
As emerging two-dimensional(2D)materials,carbides and nitrides(MXenes)could be solid solutions or organized structures made up of multi-atomic layers.With remarkable and adjustable electrical,optical,mechanical,and el...As emerging two-dimensional(2D)materials,carbides and nitrides(MXenes)could be solid solutions or organized structures made up of multi-atomic layers.With remarkable and adjustable electrical,optical,mechanical,and electrochemical characteristics,MXenes have shown great potential in brain-inspired neuromorphic computing electronics,including neuromorphic gas sensors,pressure sensors and photodetectors.This paper provides a forward-looking review of the research progress regarding MXenes in the neuromorphic sensing domain and discussed the critical challenges that need to be resolved.Key bottlenecks such as insufficient long-term stability under environmental exposure,high costs,scalability limitations in large-scale production,and mechanical mismatch in wearable integration hinder their practical deployment.Furthermore,unresolved issues like interfacial compatibility in heterostructures and energy inefficiency in neu-romorphic signal conversion demand urgent attention.The review offers insights into future research directions enhance the fundamental understanding of MXene properties and promote further integration into neuromorphic computing applications through the convergence with various emerging technologies.展开更多
The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing r...The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing resources has led to a scarcity.In order to meet the needs of an increasing number of researchers,it is imperative to facilitate efficient and flexible access to computing resources in a cloud environment.In this paper,we propose a novel quantum computing paradigm,Virtual QPU(VQPU),which addresses this issue and enhances quantum cloud throughput with guaranteed circuit fidelity.The proposal introduces three innovative concepts:(1)The integration of virtualization technology into the field of quantum computing to enhance quantum cloud throughput.(2)The introduction of an asynchronous execution of circuits methodology to improve quantum computing flexibility.(3)The development of a virtual QPU allocation scheme for quantum tasks in a cloud environment to improve circuit fidelity.The concepts have been validated through the utilization of a self-built simulated quantum cloud platform.展开更多
This comprehensive study investigates the formation and evolution of intermetallic compounds during the solidification process of magnesium alloys using advanced micro X-ray computed tomography.By analyzing both commo...This comprehensive study investigates the formation and evolution of intermetallic compounds during the solidification process of magnesium alloys using advanced micro X-ray computed tomography.By analyzing both common industrial Mg-Al-Zn alloys and a novel rare earth-containing Mg-Ni-Gd-Y alloy,we aim to characterize the nucleation,growth,and distribution of Al-Mn and eutectic intermetallics across various stages of solidification.The non destructive imaging technique employed in this research provides high-resolution,three-dimensional insights into the microstructural development,allowing for a detailed examination of the morphology,spatial arrangement,and interconnectivity of intermetallic phases.This approach overcomes limitations of traditional two-dimensional metallographic methods,offering a more comprehensive understanding of the complex three-dimensional structures formed during solidification.展开更多
New electronic devices based on the physical properties of electrically driven skyrmions are promising for logic computing and nonvolatile memory applications.However,achieving efficient and practical compute-storage ...New electronic devices based on the physical properties of electrically driven skyrmions are promising for logic computing and nonvolatile memory applications.However,achieving efficient and practical compute-storage integration remains challenging owing to the structural complexity,limited functionality,and low flexibility observed in most skyrmion-based devices.In this study,we designed a novel device architecture that integrates seven basic logic gates into a unified physical structure.Their operation can be enabled by physical mechanisms,such as spin-orbit torque,spin-transfer torque,skyrmion-edge repulsions,and skyrmion-skyrmion interactions.Furthermore,by incorporating voltage-controlled magnetic anisotropy,the device achieved multi-input capability and reconfigurability functionality.Ultralow power consumption(<1 fJ/bit per logic function)and extremely high logic density were achieved.Significantly,the compatibility of this nanotrack design with existing skyrmion racetrack memory paves the way for advanced in-memory computing in spintronic architectures.展开更多
High-entropy oxides(HEOs)have emerged as a promising class of memristive materials,characterized by entropy-stabilized crystal structures,multivalent cation coordination,and tunable defect landscapes.These intrinsic f...High-entropy oxides(HEOs)have emerged as a promising class of memristive materials,characterized by entropy-stabilized crystal structures,multivalent cation coordination,and tunable defect landscapes.These intrinsic features enable forming-free resistive switching,multilevel conductance modulation,and synaptic plasticity,making HEOs attractive for neuromorphic computing.This review outlines recent progress in HEO-based memristors across materials engineering,switching mechanisms,and synaptic emulation.Particular attention is given to vacancy migration,phase transitions,and valence-state dynamics—mechanisms that underlie the switching behaviors observed in both amorphous and crystalline systems.Their relevance to neuromorphic functions such as short-term plasticity and spike-timing-dependent learning is also examined.While encouraging results have been achieved at the device level,challenges remain in conductance precision,variability control,and scalable integration.Addressing these demands a concerted effort across materials design,interface optimization,and task-aware modeling.With such integration,HEO memristors offer a compelling pathway toward energy-efficient and adaptable brain-inspired electronics.展开更多
Memristors have emerged as a transformative technology in the realm of electronic devices,offering unique advantages such as fast switching speeds,low power consumption,and the ability to sensor-memory-compute.The app...Memristors have emerged as a transformative technology in the realm of electronic devices,offering unique advantages such as fast switching speeds,low power consumption,and the ability to sensor-memory-compute.The applications span across non-volatile memory,neuromorphic computing,hardware security,and beyond,prompting memristors to become a versatile solution for next-generation computing and data storage systems.Despite enormous potential of memristors,the transition from laboratory prototypes to large-scale applications is challenging in terms of material stability,device reproducibility,and array scalability.This review systematically explores recent advancements in high-performance memristor technologies,focusing on performance enhancement strategies through material engineering,structural design,pulse protocol optimization,and algorithm control.We provide an in-depth analysis of key performance metrics tailored to specific applications,including non-volatile memory,neuromorphic computing,and hardware security.Furthermore,we propose a co-design framework that integrates device-level optimizations with operational-level improvements,aiming to bridge the gap between theoretical models and practical implementations.展开更多
The advancement of flexible memristors has significantly promoted the development of wearable electronic for emerging neuromorphic computing applications.Inspired by in-memory computing architecture of human brain,fle...The advancement of flexible memristors has significantly promoted the development of wearable electronic for emerging neuromorphic computing applications.Inspired by in-memory computing architecture of human brain,flexible memristors exhibit great application potential in emulating artificial synapses for highefficiency and low power consumption neuromorphic computing.This paper provides comprehensive overview of flexible memristors from perspectives of development history,material system,device structure,mechanical deformation method,device performance analysis,stress simulation during deformation,and neuromorphic computing applications.The recent advances in flexible electronics are summarized,including single device,device array and integration.The challenges and future perspectives of flexible memristor for neuromorphic computing are discussed deeply,paving the way for constructing wearable smart electronics and applications in large-scale neuromorphic computing and high-order intelligent robotics.展开更多
Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce different...Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.展开更多
The intelligent environmental sensing systems are quickly transforming the sparse and retrospective monitoring to dense and decision-oriented environmental intelligence.This review brings together the manner in which ...The intelligent environmental sensing systems are quickly transforming the sparse and retrospective monitoring to dense and decision-oriented environmental intelligence.This review brings together the manner in which integration of Internet of Things(IoT)sensing,edge computing,and real-time analytics facilitates timely detection,interpretation,and prediction of the environmental conditions across the applications,such as urban air quality,watershed and coastal surveillance,industrial safety,agriculture,and disaster response.We define end-to-end architectural patterns to organize devices,edge nodes,and cloud services to satisfy latency,reliability,bandwidth,and governance constraints with emphasis on event-time processing,adaptive offloading,and hierarchical aggregation.Then we look at sensing and infrastructure foundations,emphasizing the effects of sensor modality and power autonomy,connectivity,and the practices of calibration on the practicable analytics and eventual plausibility.It is on this basis that we examine real-time analytics pipelines and Artificial Intelligence(AI)techniques to preprocess,sensor combine,anomaly detect,and short-horizon forecast,with a focus on edge-deployable models,quantification of uncertainties,and query resistance to drift and domain shift.Lastly,we address the realities of deployment that condition operational success,such as lifecycle engineering,provenance-aware data management,security and privacy risks,ethical governance,and evaluation methodologies,which place end-to-end latency and field generalization as a priority.This review offers cohesion to algorithmic capabilities and systems engineering and governance to define an overall framework,show open areas of research directions,and provide practical recommendations on how to design trustworthy,scalable,and sustainable environmental monitoring systems.展开更多
基金supported in part by the National Natural Science Foundation of China under Grant 52307134the Fundamental Research Funds for the Central Universities(xzy012025022)。
文摘Active distribution network(ADN)planning is crucial for achieving a cost-effective transition to modern power systems,yet it poses significant challenges as the system scale increases.The advent of quantum computing offers a transformative approach to solve ADN planning.To fully leverage the potential of quantum computing,this paper proposes a photonic quantum acceleration algorithm.First,a quantum-accelerated framework for ADN planning is proposed on the basis of coherent photonic quantum computers.The ADN planning model is then formulated and decomposed into discrete master problems and continuous subproblems to facilitate the quantum optimization process.The photonic quantum-embedded adaptive alternating direction method of multipliers(PQA-ADMM)algorithm is subsequently proposed to equivalently map the discrete master problem onto a quantum-interpretable model,enabling its deployment on a photonic quantum computer.Finally,a comparative analysis with various solvers,including Gurobi,demonstrates that the proposed PQA-ADMM algorithm achieves significant speedup on the modified IEEE 33-node and IEEE 123-node systems,highlighting its effectiveness.
文摘The advent of quantum computing poses a significant challenge to traditional cryptographic protocols,particularly those used in SecureMultiparty Computation(MPC),a fundamental cryptographic primitive for privacypreserving computation.Classical MPC relies on cryptographic techniques such as homomorphic encryption,secret sharing,and oblivious transfer,which may become vulnerable in the post-quantum era due to the computational power of quantum adversaries.This study presents a review of 140 peer-reviewed articles published between 2000 and 2025 that used different databases like MDPI,IEEE Explore,Springer,and Elsevier,examining the applications,types,and security issues with the solution of Quantum computing in different fields.This review explores the impact of quantum computing on MPC security,assesses emerging quantum-resistant MPC protocols,and examines hybrid classicalquantum approaches aimed at mitigating quantum threats.We analyze the role of Quantum Key Distribution(QKD),post-quantum cryptography(PQC),and quantum homomorphic encryption in securing multiparty computations.Additionally,we discuss the challenges of scalability,computational efficiency,and practical deployment of quantumsecure MPC frameworks in real-world applications such as privacy-preserving AI,secure blockchain transactions,and confidential data analysis.This review provides insights into the future research directions and open challenges in ensuring secure,scalable,and quantum-resistant multiparty computation.
基金appreciation to the Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R384)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.
基金supported by the National Key Research and Development Program of China under Grant 2022YFB3608300in part by the National Nature Science Foundation of China(NSFC)under Grants 62404050,U2341218,62574056,62204052。
文摘Organic electrochemical transistor(OECT)devices demonstrate great promising potential for reservoir computing(RC)systems,but their lack of tunable dynamic characteristics limits their application in multi-temporal scale tasks.In this study,we report an OECT-based neuromorphic device with tunable relaxation time(τ)by introducing an additional vertical back-gate electrode into a planar structure.The dual-gate design enablesτreconfiguration from 93 to 541 ms.The tunable relaxation behaviors can be attributed to the combined effects of planar-gate induced electrochemical doping and back-gateinduced electrostatic coupling,as verified by electrochemical impedance spectroscopy analysis.Furthermore,we used theτ-tunable OECT devices as physical reservoirs in the RC system for intelligent driving trajectory prediction,achieving a significant improvement in prediction accuracy from below 69%to 99%.The results demonstrate that theτ-tunable OECT shows a promising candidate for multi-temporal scale neuromorphic computing applications.
基金supported by the National Natural Science Foundation of China(Grants No.92364204 and 62204219)the open research fund of Suzhou Laboratory(Grants No.SZLAB-1208-2024-TS012)+1 种基金Major Program of Natural Science Foundation of Zhejiang Province(Grants No.LDT23F0401)Zhejiang Province Introduces and Cultivates Leading Innovation and Entrepreneurship Teams(Grants No.2023R01011)。
文摘The deceleration of Moore's law and the energy–latency drawbacks of the von Neumann bottleneck have heightened the pursuit for beyond-CMOS designs that integrate memory and compute.Self-rectifying memristors(SRMs)have emerged as promising building blocks for high-performance,low-power systems by combining resistive switching with intrinsic diode-like behavior.Their unidirectional conduction inhibits sneak-path currents in crossbar arrays devoid of external selectors,while nonlinear I–V characteristics,adjustable conductance states,low operating voltages,and rapid switching facilitate efficient vector–matrix operations,neuromorphic plasticity,and hardware security primitives.This review synthesizes the working mechanisms of SRMs,surveys material,and structural strategies and compares device metrics relevant to array-scale deployment(rectification ratio,nonlinearity,endurance,retention,variability,and operating voltage).We assess SRM-enabled in-memory computing and neuromorphic applications,as well as security functions such as physical unclonable functions and reconfigurable cryptographic primitives.Integration pathways toward CMOS compatibility are analyzed,including back-end-of-line thermal budgets,uniformity,write disturb mitigation,and reliability.Finally,we outline key challenges and opportunities:materials/architecture co-design,precision analog training,stochasticity control/exploitation,3D stacking,and standardized benchmarking that can accelerate large-scale SRM adoption.Through the use of specialized materials and structural optimization,SRMs are set to provide selector-free,densely integrated,and energy-efficient hardware for future information processing.
基金supported by Key Science and Technology Program of Henan Province,China(Grant Nos.242102210147,242102210027)Fujian Province Young and Middle aged Teacher Education Research Project(Science and Technology Category)(No.JZ240101)(Corresponding author:Dong Yuan).
文摘Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.
文摘The integration of large-scale foundation models(e.g.,GPT series and AlphaFold)into oncology is fundamentally transforming both research methodologies and clinical practices,driven by unprecedented advancements in computational power.This review synthesizes recent progress in the application of large language models to core oncological tasks,including medical imaging analysis,genomic interpretation,and personalized treatment planning.Underpinned by advanced computational infrastructures,such as graphics processing unit/tensor processing unit clusters,heterogeneous computing,and cloud platforms,these models enable superior representation learning and generalization across multimodal data sources.This review examines how these infrastructures overcome key bottlenecks in intelligent oncology through scalable optimization strategies,including mixed-precision training,memory optimization,and heterogeneous computing.Alongside these technical advancements,the review explores pressing challenges,such as data heterogeneity,limited model interpretability,regulatory uncertainties,and the environmental impact of artificial intelligence(AI)systems.Special emphasis is placed on emerging solutions,encompassing green AI and edge computing,which offer promising approaches for low-resource deployment scenarios.Additionally,the review highlights the critical role of interdisciplinary collaboration among oncology,computer science,ethics,and policy to ensure that AI systems are not only powerful but also transparent,safe,and clinically relevant.Finally,the review outlines potential avenues for future research aimed at developing robust,scalable,and human-centered frameworks for intelligent oncology.
基金supported by the National Natural Science Foundation of China(62202215)Liaoning Province Applied Basic Research Program(Youth Special Project,2023JH2/101600038)+4 种基金Shenyang Youth Science and Technology Innovation Talent Support Program(RC220458)Guangxuan Program of Shenyang Ligong University(SYLUGXRC202216)the Basic Research Special Funds for Undergraduate Universities in Liaoning Province(LJ212410144067)the Natural Science Foundation of Liaoning Province(2024-MS-113)the science and technology funds from Liaoning Education Department(LJKZ0242).
文摘In the field of edge computing,achieving low-latency computational task offloading with limited resources is a critical research challenge,particularly in resource-constrained and latency-sensitive vehicular network environments where rapid response is mandatory for safety-critical applications.In scenarios where edge servers are sparsely deployed,the lack of coordination and information sharing often leads to load imbalance,thereby increasing system latency.Furthermore,in regions without edge server coverage,tasks must be processed locally,which further exacerbates latency issues.To address these challenges,we propose a novel and efficient Deep Reinforcement Learning(DRL)-based approach aimed at minimizing average task latency.The proposed method incorporates three offloading strategies:local computation,direct offloading to the edge server in local region,and device-to-device(D2D)-assisted offloading to edge servers in other regions.We formulate the task offloading process as a complex latency minimization optimization problem.To solve it,we propose an advanced algorithm based on the Dueling Double Deep Q-Network(D3QN)architecture and incorporating the Prioritized Experience Replay(PER)mechanism.Experimental results demonstrate that,compared with existing offloading algorithms,the proposed method significantly reduces average task latency,enhances user experience,and offers an effective strategy for latency optimization in future edge computing systems under dynamic workloads.
基金supported by the National Key Research and Development Program of China(No.2022YFF0607802)。
文摘Rotational computed laminography(CL)has broad application potential in three-dimensional imaging of plate-like objects because it only requires X-rays to pass through the tested object in the thickness direction during the imaging process.In this study,a rectangular cross-section field-of-view rotational CL(RC-CL)is proposed for circuit board imaging.Compared to other rotational CL systems,the field of view is the largest and most suitable for rectangular circuit boards.Meanwhile,as the imaging geometry of RC-CL is significantly different from that of cone-beam CT,the Feldkamp-Davis-Kress(FDK)reconstruction algorithm cannot be used directly.However,transferring the projection data to fit into the CBCT geometry using two-dimensional interpolation introduces interpolation errors.Therefore,an FDK-type analytical reconstruction algorithm applicable to RC-CL was developed.The effectiveness of the method was validated through numerical experiments,and the influence of the tilt angle on the reconstruction results was analyzed.Finally,the RC-CL technique was applied to real defect detection research on circuit boards.
基金supported and funded by theDeanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-DDRSP2503).
文摘In recent years,fog computing has become an important environment for dealing with the Internet of Things.Fog computing was developed to handle large-scale big data by scheduling tasks via cloud computing.Task scheduling is crucial for efficiently handling IoT user requests,thereby improving system performance,cost,and energy consumption across nodes in cloud computing.With the large amount of data and user requests,achieving the optimal solution to the task scheduling problem is challenging,particularly in terms of cost and energy efficiency.In this paper,we develop novel strategies to save energy consumption across nodes in fog computing when users execute tasks through the least-cost paths.Task scheduling is developed using modified artificial ecosystem optimization(AEO),combined with negative swarm operators,Salp Swarm Algorithm(SSA),in order to competitively optimize their capabilities during the exploitation phase of the optimal search process.In addition,the proposed strategy,Enhancement Artificial Ecosystem Optimization Salp Swarm Algorithm(EAEOSSA),attempts to find the most suitable solution.The optimization that combines cost and energy for multi-objective task scheduling optimization problems.The backpack problem is also added to improve both cost and energy in the iFogSim implementation as well.A comparison was made between the proposed strategy and other strategies in terms of time,cost,energy,and productivity.Experimental results showed that the proposed strategy improved energy consumption,cost,and time over other algorithms.Simulation results demonstrate that the proposed algorithm increases the average cost,average energy consumption,and mean service time in most scenarios,with average reductions of up to 21.15%in cost and 25.8%in energy consumption.
基金funded by the State Grid Corporation Science and Technology Project“Research and Application of Key Technologies for Integrated Sensing and Computing for Intelligent Operation of Power Grid”(Grant No.5700-202318596A-3-2-ZN).
文摘With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on computing latency and resource supply efficiency.Multi-access edge computing technology deploys cloud computing capabilities at the network edge;constructs distributed computing nodes and multi-access systems and offers infrastructure support for services with low latency and high reliability.Existing research relies on a strong assumption that the environmental state is fully observable and fails to thoroughly consider the continuous time-varying features of edge server load fluctuations,leading to insufficient adaptability of the model in a heterogeneous dynamic environment.Thus,this paper establishes a framework for end-edge collaborative task offloading based on a partially observable Markov decision-making process(POMDP)and proposes a method for end-edge collaborative task offloading in heterogeneous scenarios.It achieves time-series modeling of the historical load characteristics of edge servers and endows the agent with the ability to be aware of the load in dynamic environmental states.Moreover,by dynamically assessing the exploration value of historical trajectories in the central trajectory pool and adjusting the sample weight distribution,directional exploration and strategy optimization of high-value trajectories are realized.Experimental results indicate that the proposed method exhibits distinct advantages compared with existing methods in terms of average delay and task failure rate and also verifies the method’s robustness in a dynamic environment.
基金supported by the NSFC(12474071)Natural Science Foundation of Shandong Province(ZR2024YQ051,ZR2025QB50)+6 种基金Guangdong Basic and Applied Basic Research Foundation(2025A1515011191)the Shanghai Sailing Program(23YF1402200,23YF1402400)funded by Basic Research Program of Jiangsu(BK20240424)Open Research Fund of State Key Laboratory of Crystal Materials(KF2406)Taishan Scholar Foundation of Shandong Province(tsqn202408006,tsqn202507058)Young Talent of Lifting engineering for Science and Technology in Shandong,China(SDAST2024QTB002)the Qilu Young Scholar Program of Shandong University。
文摘As emerging two-dimensional(2D)materials,carbides and nitrides(MXenes)could be solid solutions or organized structures made up of multi-atomic layers.With remarkable and adjustable electrical,optical,mechanical,and electrochemical characteristics,MXenes have shown great potential in brain-inspired neuromorphic computing electronics,including neuromorphic gas sensors,pressure sensors and photodetectors.This paper provides a forward-looking review of the research progress regarding MXenes in the neuromorphic sensing domain and discussed the critical challenges that need to be resolved.Key bottlenecks such as insufficient long-term stability under environmental exposure,high costs,scalability limitations in large-scale production,and mechanical mismatch in wearable integration hinder their practical deployment.Furthermore,unresolved issues like interfacial compatibility in heterostructures and energy inefficiency in neu-romorphic signal conversion demand urgent attention.The review offers insights into future research directions enhance the fundamental understanding of MXene properties and promote further integration into neuromorphic computing applications through the convergence with various emerging technologies.
文摘The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing resources has led to a scarcity.In order to meet the needs of an increasing number of researchers,it is imperative to facilitate efficient and flexible access to computing resources in a cloud environment.In this paper,we propose a novel quantum computing paradigm,Virtual QPU(VQPU),which addresses this issue and enhances quantum cloud throughput with guaranteed circuit fidelity.The proposal introduces three innovative concepts:(1)The integration of virtualization technology into the field of quantum computing to enhance quantum cloud throughput.(2)The introduction of an asynchronous execution of circuits methodology to improve quantum computing flexibility.(3)The development of a virtual QPU allocation scheme for quantum tasks in a cloud environment to improve circuit fidelity.The concepts have been validated through the utilization of a self-built simulated quantum cloud platform.
基金Project(2023YFB4606200)supported by the National Key Research and Development Program of ChinaProject(2023-SSRF-HZ-503114-2)supported by Shanghai Synchrotron Radiation Facility,Instrument BL16U2,China。
文摘This comprehensive study investigates the formation and evolution of intermetallic compounds during the solidification process of magnesium alloys using advanced micro X-ray computed tomography.By analyzing both common industrial Mg-Al-Zn alloys and a novel rare earth-containing Mg-Ni-Gd-Y alloy,we aim to characterize the nucleation,growth,and distribution of Al-Mn and eutectic intermetallics across various stages of solidification.The non destructive imaging technique employed in this research provides high-resolution,three-dimensional insights into the microstructural development,allowing for a detailed examination of the morphology,spatial arrangement,and interconnectivity of intermetallic phases.This approach overcomes limitations of traditional two-dimensional metallographic methods,offering a more comprehensive understanding of the complex three-dimensional structures formed during solidification.
基金support from the National Natural Science Foundation of China (Grant No.12474101)support from the National Natural Science Foundation of China (Grant Nos.52272202 and W2421027)support from the National Natural Science Foundation of China (Grant No.52501307)。
文摘New electronic devices based on the physical properties of electrically driven skyrmions are promising for logic computing and nonvolatile memory applications.However,achieving efficient and practical compute-storage integration remains challenging owing to the structural complexity,limited functionality,and low flexibility observed in most skyrmion-based devices.In this study,we designed a novel device architecture that integrates seven basic logic gates into a unified physical structure.Their operation can be enabled by physical mechanisms,such as spin-orbit torque,spin-transfer torque,skyrmion-edge repulsions,and skyrmion-skyrmion interactions.Furthermore,by incorporating voltage-controlled magnetic anisotropy,the device achieved multi-input capability and reconfigurability functionality.Ultralow power consumption(<1 fJ/bit per logic function)and extremely high logic density were achieved.Significantly,the compatibility of this nanotrack design with existing skyrmion racetrack memory paves the way for advanced in-memory computing in spintronic architectures.
基金financially supported by the National Natural Science Foundation of China(Grant No.12172093)the Guangdong Basic and Applied Basic Research Foundation(Grant No.2021A1515012607)。
文摘High-entropy oxides(HEOs)have emerged as a promising class of memristive materials,characterized by entropy-stabilized crystal structures,multivalent cation coordination,and tunable defect landscapes.These intrinsic features enable forming-free resistive switching,multilevel conductance modulation,and synaptic plasticity,making HEOs attractive for neuromorphic computing.This review outlines recent progress in HEO-based memristors across materials engineering,switching mechanisms,and synaptic emulation.Particular attention is given to vacancy migration,phase transitions,and valence-state dynamics—mechanisms that underlie the switching behaviors observed in both amorphous and crystalline systems.Their relevance to neuromorphic functions such as short-term plasticity and spike-timing-dependent learning is also examined.While encouraging results have been achieved at the device level,challenges remain in conductance precision,variability control,and scalable integration.Addressing these demands a concerted effort across materials design,interface optimization,and task-aware modeling.With such integration,HEO memristors offer a compelling pathway toward energy-efficient and adaptable brain-inspired electronics.
基金supported by the National Key R&D Project from the Minister of Science and Technology(2024YFA1211500)the National Natural Science Foundation of China(Grant Nos.62304130,62405158 and 62574123)+1 种基金the Shanghai youth science and technology star project(24QA2702800)Shanghai Key Laboratory of Chips and Systems for Intelligent Connected Vehicle。
文摘Memristors have emerged as a transformative technology in the realm of electronic devices,offering unique advantages such as fast switching speeds,low power consumption,and the ability to sensor-memory-compute.The applications span across non-volatile memory,neuromorphic computing,hardware security,and beyond,prompting memristors to become a versatile solution for next-generation computing and data storage systems.Despite enormous potential of memristors,the transition from laboratory prototypes to large-scale applications is challenging in terms of material stability,device reproducibility,and array scalability.This review systematically explores recent advancements in high-performance memristor technologies,focusing on performance enhancement strategies through material engineering,structural design,pulse protocol optimization,and algorithm control.We provide an in-depth analysis of key performance metrics tailored to specific applications,including non-volatile memory,neuromorphic computing,and hardware security.Furthermore,we propose a co-design framework that integrates device-level optimizations with operational-level improvements,aiming to bridge the gap between theoretical models and practical implementations.
基金supported by the NSFC(12474071)Natural Science Foundation of Shandong Province(ZR2024YQ051)+5 种基金Open Research Fund of State Key Laboratory of Materials for Integrated Circuits(SKLJC-K2024-12)the Shanghai Sailing Program(23YF1402200,23YF1402400)Natural Science Foundation of Jiangsu Province(BK20240424)Taishan Scholar Foundation of Shandong Province(tsqn202408006)Young Talent of Lifting engineering for Science and Technology in Shandong,China(SDAST2024QTB002)the Qilu Young Scholar Program of Shandong University.
文摘The advancement of flexible memristors has significantly promoted the development of wearable electronic for emerging neuromorphic computing applications.Inspired by in-memory computing architecture of human brain,flexible memristors exhibit great application potential in emulating artificial synapses for highefficiency and low power consumption neuromorphic computing.This paper provides comprehensive overview of flexible memristors from perspectives of development history,material system,device structure,mechanical deformation method,device performance analysis,stress simulation during deformation,and neuromorphic computing applications.The recent advances in flexible electronics are summarized,including single device,device array and integration.The challenges and future perspectives of flexible memristor for neuromorphic computing are discussed deeply,paving the way for constructing wearable smart electronics and applications in large-scale neuromorphic computing and high-order intelligent robotics.
基金funded by National Research Council of Thailand(contract No.N42A671047).
文摘Physics-informed neural networks(PINNs)have emerged as a promising class of scientific machine learning techniques that integrate governing physical laws into neural network training.Their ability to enforce differential equations,constitutive relations,and boundary conditions within the loss function provides a physically grounded alternative to traditional data-driven models,particularly for solid and structural mechanics,where data are often limited or noisy.This review offers a comprehensive assessment of recent developments in PINNs,combining bibliometric analysis,theoretical foundations,application-oriented insights,and methodological innovations.A biblio-metric survey indicates a rapid increase in publications on PINNs since 2018,with prominent research clusters focused on numerical methods,structural analysis,and forecasting.Building upon this trend,the review consolidates advance-ments across five principal application domains,including forward structural analysis,inverse modeling and parameter identification,structural and topology optimization,assessment of structural integrity,and manufacturing processes.These applications are propelled by substantial methodological advancements,encompassing rigorous enforcement of boundary conditions,modified loss functions,adaptive training,domain decomposition strategies,multi-fidelity and transfer learning approaches,as well as hybrid finite element–PINN integration.These advances address recurring challenges in solid mechanics,such as high-order governing equations,material heterogeneity,complex geometries,localized phenomena,and limited experimental data.Despite remaining challenges in computational cost,scalability,and experimental validation,PINNs are increasingly evolving into specialized,physics-aware tools for practical solid and structural mechanics applications.
基金supported by Jiangxi Polytechnic Institute Key Research Topics in Educational Reform 2025-JGJG-07.
文摘The intelligent environmental sensing systems are quickly transforming the sparse and retrospective monitoring to dense and decision-oriented environmental intelligence.This review brings together the manner in which integration of Internet of Things(IoT)sensing,edge computing,and real-time analytics facilitates timely detection,interpretation,and prediction of the environmental conditions across the applications,such as urban air quality,watershed and coastal surveillance,industrial safety,agriculture,and disaster response.We define end-to-end architectural patterns to organize devices,edge nodes,and cloud services to satisfy latency,reliability,bandwidth,and governance constraints with emphasis on event-time processing,adaptive offloading,and hierarchical aggregation.Then we look at sensing and infrastructure foundations,emphasizing the effects of sensor modality and power autonomy,connectivity,and the practices of calibration on the practicable analytics and eventual plausibility.It is on this basis that we examine real-time analytics pipelines and Artificial Intelligence(AI)techniques to preprocess,sensor combine,anomaly detect,and short-horizon forecast,with a focus on edge-deployable models,quantification of uncertainties,and query resistance to drift and domain shift.Lastly,we address the realities of deployment that condition operational success,such as lifecycle engineering,provenance-aware data management,security and privacy risks,ethical governance,and evaluation methodologies,which place end-to-end latency and field generalization as a priority.This review offers cohesion to algorithmic capabilities and systems engineering and governance to define an overall framework,show open areas of research directions,and provide practical recommendations on how to design trustworthy,scalable,and sustainable environmental monitoring systems.