Numerical simulation is dominant in solving partial differential equations(PDEs),but balancing fine-grained grids with low computational costs is challenging.Recently,solving PDEs with neural networks(NNs)has gained i...Numerical simulation is dominant in solving partial differential equations(PDEs),but balancing fine-grained grids with low computational costs is challenging.Recently,solving PDEs with neural networks(NNs)has gained interest,yet cost-effectiveness and high accuracy remain a challenge.This work introduces a novel paradigm for solving PDEs,called multi-scale neural computing(MSNC),considering spectral bias of NNs and local approximation properties in the finite difference method(FDM).The MSNC decomposes the solution with a NN for efficient capture of global scale and the FDM for detailed description of local scale,aiming to balance costs and accuracy.Demonstrated advantages include higher accuracy(10 times for 1D PDEs,20 times for 2D PDEs)and lower costs(4 times for 1D PDEs,16 times for 2D PDEs)than the standard FDM.The MSNC also exhibits stable convergence and rigorous boundary condition satisfaction,showcasing the potential for hybrid of NN and numerical method.展开更多
With the rapid growth of socialmedia,the spread of fake news has become a growing problem,misleading the public and causing significant harm.As social media content is often composed of both images and text,the use of...With the rapid growth of socialmedia,the spread of fake news has become a growing problem,misleading the public and causing significant harm.As social media content is often composed of both images and text,the use of multimodal approaches for fake news detection has gained significant attention.To solve the problems existing in previous multi-modal fake news detection algorithms,such as insufficient feature extraction and insufficient use of semantic relations between modes,this paper proposes the MFFFND-Co(Multimodal Feature Fusion Fake News Detection with Co-Attention Block)model.First,the model deeply explores the textual content,image content,and frequency domain features.Then,it employs a Co-Attention mechanism for cross-modal fusion.Additionally,a semantic consistency detectionmodule is designed to quantify semantic deviations,thereby enhancing the performance of fake news detection.Experimentally verified on two commonly used datasets,Twitter and Weibo,the model achieved F1 scores of 90.0% and 94.0%,respectively,significantly outperforming the pre-modified MFFFND(Multimodal Feature Fusion Fake News Detection with Attention Block)model and surpassing other baseline models.This improves the accuracy of detecting fake information in artificial intelligence detection and engineering software detection.展开更多
Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in us...Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.展开更多
Because of the challenge of compounding lightweight,high-strength Ti/Al alloys due to their considerable disparity in properties,Al 6063 as intermediate layer was proposed to fabricate TC4/Al 6063/Al 7075 three-layer ...Because of the challenge of compounding lightweight,high-strength Ti/Al alloys due to their considerable disparity in properties,Al 6063 as intermediate layer was proposed to fabricate TC4/Al 6063/Al 7075 three-layer composite plate by explosive welding.The microscopic properties of each bonding interface were elucidated through field emission scanning electron microscope and electron backscattered diffraction(EBSD).A methodology combining finite element method-smoothed particle hydrodynamics(FEM-SPH)and molecular dynamics(MD)was proposed for the analysis of the forming and evolution characteristics of explosive welding interfaces at multi-scale.The results demonstrate that the bonding interface morphologies of TC4/Al 6063 and Al 6063/Al 7075 exhibit a flat and wavy configuration,without discernible defects or cracks.The phenomenon of grain refinement is observed in the vicinity of the two bonding interfaces.Furthermore,the degree of plastic deformation of TC4 and Al 7075 is more pronounced than that of Al 6063 in the intermediate layer.The interface morphology characteristics obtained by FEM-SPH simulation exhibit a high degree of similarity to the experimental results.MD simulations reveal that the diffusion of interfacial elements predominantly occurs during the unloading phase,and the simulated thickness of interfacial diffusion aligns well with experimental outcomes.The introduction of intermediate layer in the explosive welding process can effectively produce high-quality titanium/aluminum alloy composite plates.Furthermore,this approach offers a multi-scale simulation strategy for the study of explosive welding bonding interfaces.展开更多
Optoelectronic memristor is generating growing research interest for high efficient computing and sensing-memory applications.In this work,an optoelectronic memristor with Au/a-C:Te/Pt structure is developed.Synaptic ...Optoelectronic memristor is generating growing research interest for high efficient computing and sensing-memory applications.In this work,an optoelectronic memristor with Au/a-C:Te/Pt structure is developed.Synaptic functions,i.e.,excita-tory post-synaptic current and pair-pulse facilitation are successfully mimicked with the memristor under electrical and optical stimulations.More importantly,the device exhibited distinguishable response currents by adjusting 4-bit input electrical/opti-cal signals.A multi-mode reservoir computing(RC)system is constructed with the optoelectronic memristors to emulate human tactile-visual fusion recognition and an accuracy of 98.7%is achieved.The optoelectronic memristor provides potential for developing multi-mode RC system.展开更多
Prediction of production decline and evaluation of the adsorbed/free gas ratio are critical for determining the lifespan and production status of shale gas wells.Traditional production prediction methods have some sho...Prediction of production decline and evaluation of the adsorbed/free gas ratio are critical for determining the lifespan and production status of shale gas wells.Traditional production prediction methods have some shortcomings because of the low permeability and tightness of shale,complex gas flow behavior of multi-scale gas transport regions and multiple gas transport mechanism superpositions,and complex and variable production regimes of shale gas wells.Recent research has demonstrated the existence of a multi-stage isotope fractionation phenomenon during shale gas production,with the fractionation characteristics of each stage associated with the pore structure,gas in place(GIP),adsorption/desorption,and gas production process.This study presents a new approach for estimating shale gas well production and evaluating the adsorbed/free gas ratio throughout production using isotope fractionation techniques.A reservoir-scale carbon isotope fractionation(CIF)model applicable to the production process of shale gas wells was developed for the first time in this research.In contrast to the traditional model,this model improves production prediction accuracy by simultaneously fitting the gas production rate and δ^(13)C_(1) data and provides a new evaluation method of the adsorbed/free gas ratio during shale gas production.The results indicate that the diffusion and adsorption/desorption properties of rock,bottom-hole flowing pressure(BHP)of gas well,and multi-scale gas transport regions of the reservoir all affect isotope fractionation,with the diffusion and adsorption/desorption parameters of rock having the greatest effect on isotope fractionation being D∗/D,PL,VL,α,and others in that order.We effectively tested the universality of the four-stage isotope fractionation feature and revealed a unique isotope fractionation mechanism caused by the superimposed coupling of multi-scale gas transport regions during shale gas well production.Finally,we applied the established CIF model to a shale gas well in the Sichuan Basin,China,and calculated the estimated ultimate recovery(EUR)of the well to be 3.33×10^(8) m^(3);the adsorbed gas ratio during shale gas production was 1.65%,10.03%,and 23.44%in the first,fifth,and tenth years,respectively.The findings are significant for understanding the isotope fractionation mechanism during natural gas transport in complex systems and for formulating and optimizing unconventional natural gas development strategies.展开更多
As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the...As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the task offloading strategies by interacting with the entities. In actual application scenarios, users of edge computing are always changing dynamically. However, the existing task offloading strategies cannot be applied to such dynamic scenarios. To solve this problem, we propose a novel dynamic task offloading framework for distributed edge computing, leveraging the potential of meta-reinforcement learning (MRL). Our approach formulates a multi-objective optimization problem aimed at minimizing both delay and energy consumption. We model the task offloading strategy using a directed acyclic graph (DAG). Furthermore, we propose a distributed edge computing adaptive task offloading algorithm rooted in MRL. This algorithm integrates multiple Markov decision processes (MDP) with a sequence-to-sequence (seq2seq) network, enabling it to learn and adapt task offloading strategies responsively across diverse network environments. To achieve joint optimization of delay and energy consumption, we incorporate the non-dominated sorting genetic algorithm II (NSGA-II) into our framework. Simulation results demonstrate the superiority of our proposed solution, achieving a 21% reduction in time delay and a 19% decrease in energy consumption compared to alternative task offloading schemes. Moreover, our scheme exhibits remarkable adaptability, responding swiftly to changes in various network environments.展开更多
In recent years,gait-based emotion recognition has been widely applied in the field of computer vision.However,existing gait emotion recognition methods typically rely on complete human skeleton data,and their accurac...In recent years,gait-based emotion recognition has been widely applied in the field of computer vision.However,existing gait emotion recognition methods typically rely on complete human skeleton data,and their accuracy significantly declines when the data is occluded.To enhance the accuracy of gait emotion recognition under occlusion,this paper proposes a Multi-scale Suppression Graph ConvolutionalNetwork(MS-GCN).TheMS-GCN consists of three main components:Joint Interpolation Module(JI Moudle),Multi-scale Temporal Convolution Network(MS-TCN),and Suppression Graph Convolutional Network(SGCN).The JI Module completes the spatially occluded skeletal joints using the(K-Nearest Neighbors)KNN interpolation method.The MS-TCN employs convolutional kernels of various sizes to comprehensively capture the emotional information embedded in the gait,compensating for the temporal occlusion of gait information.The SGCN extracts more non-prominent human gait features by suppressing the extraction of key body part features,thereby reducing the negative impact of occlusion on emotion recognition results.The proposed method is evaluated on two comprehensive datasets:Emotion-Gait,containing 4227 real gaits from sources like BML,ICT-Pollick,and ELMD,and 1000 synthetic gaits generated using STEP-Gen technology,and ELMB,consisting of 3924 gaits,with 1835 labeled with emotions such as“Happy,”“Sad,”“Angry,”and“Neutral.”On the standard datasets Emotion-Gait and ELMB,the proposed method achieved accuracies of 0.900 and 0.896,respectively,attaining performance comparable to other state-ofthe-artmethods.Furthermore,on occlusion datasets,the proposedmethod significantly mitigates the performance degradation caused by occlusion compared to other methods,the accuracy is significantly higher than that of other methods.展开更多
The rise of large-scale artificial intelligence(AI)models,such as ChatGPT,Deep-Seek,and autonomous vehicle systems,has significantly advanced the boundaries of AI,enabling highly complex tasks in natural language proc...The rise of large-scale artificial intelligence(AI)models,such as ChatGPT,Deep-Seek,and autonomous vehicle systems,has significantly advanced the boundaries of AI,enabling highly complex tasks in natural language processing,image recognition,and real-time decisionmaking.However,these models demand immense computational power and are often centralized,relying on cloud-based architectures with inherent limitations in latency,privacy,and energy efficiency.To address these challenges and bring AI closer to real-world applications,such as wearable health monitoring,robotics,and immersive virtual environments,innovative hardware solutions are urgently needed.This work introduces a near-sensor edge computing(NSEC)system,built on a bilayer AlN/Si waveguide platform,to provide real-time,energy-efficient AI capabilities at the edge.Leveraging the electro-optic properties of AlN microring resonators for photonic feature extraction,coupled with Si-based thermo-optic Mach-Zehnder interferometers for neural network computations,the system represents a transformative approach to AI hardware design.Demonstrated through multimodal gesture and gait analysis,the NSEC system achieves high classification accuracies of 96.77%for gestures and 98.31%for gaits,ultra-low latency(<10 ns),and minimal energy consumption(<0.34 pJ).This groundbreaking system bridges the gap between AI models and real-world applications,enabling efficient,privacy-preserving AI solutions for healthcare,robotics,and next-generation human-machine interfaces,marking a pivotal advancement in edge computing and AI deployment.展开更多
To address the increasing demand for massive data storage and processing,brain-inspired neuromorphic comput-ing systems based on artificial synaptic devices have been actively developed in recent years.Among the vario...To address the increasing demand for massive data storage and processing,brain-inspired neuromorphic comput-ing systems based on artificial synaptic devices have been actively developed in recent years.Among the various materials inves-tigated for the fabrication of synaptic devices,silicon carbide(SiC)has emerged as a preferred choices due to its high electron mobility,superior thermal conductivity,and excellent thermal stability,which exhibits promising potential for neuromorphic applications in harsh environments.In this review,the recent progress in SiC-based synaptic devices is summarized.Firstly,an in-depth discussion is conducted regarding the categories,working mechanisms,and structural designs of these devices.Subse-quently,several application scenarios for SiC-based synaptic devices are presented.Finally,a few perspectives and directions for their future development are outlined.展开更多
The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method f...The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method for infrared and visible image fusion is proposed.The encoder designed according to the optimization objective consists of a base encoder and a detail encoder,which is used to extract low-frequency and high-frequency information from the image.This extraction may lead to some information not being captured,so a compensation encoder is proposed to supplement the missing information.Multi-scale decomposition is also employed to extract image features more comprehensively.The decoder combines low-frequency,high-frequency and supplementary information to obtain multi-scale features.Subsequently,the attention strategy and fusion module are introduced to perform multi-scale fusion for image reconstruction.Experimental results on three datasets show that the fused images generated by this network effectively retain salient targets while being more consistent with human visual perception.展开更多
The rapid advent in artificial intelligence and big data has revolutionized the dynamic requirement in the demands of the computing resource for executing specific tasks in the cloud environment.The process of achievi...The rapid advent in artificial intelligence and big data has revolutionized the dynamic requirement in the demands of the computing resource for executing specific tasks in the cloud environment.The process of achieving autonomic resource management is identified to be a herculean task due to its huge distributed and heterogeneous environment.Moreover,the cloud network needs to provide autonomic resource management and deliver potential services to the clients by complying with the requirements of Quality-of-Service(QoS)without impacting the Service Level Agreements(SLAs).However,the existing autonomic cloud resource managing frameworks are not capable in handling the resources of the cloud with its dynamic requirements.In this paper,Coot Bird Behavior Model-based Workload Aware Autonomic Resource Management Scheme(CBBM-WARMS)is proposed for handling the dynamic requirements of cloud resources through the estimation of workload that need to be policed by the cloud environment.This CBBM-WARMS initially adopted the algorithm of adaptive density peak clustering for workloads clustering of the cloud.Then,it utilized the fuzzy logic during the process of workload scheduling for achieving the determining the availability of cloud resources.It further used CBBM for potential Virtual Machine(VM)deployment that attributes towards the provision of optimal resources.It is proposed with the capability of achieving optimal QoS with minimized time,energy consumption,SLA cost and SLA violation.The experimental validation of the proposed CBBMWARMS confirms minimized SLA cost of 19.21%and reduced SLA violation rate of 18.74%,better than the compared autonomic cloud resource managing frameworks.展开更多
Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power sta...Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power stations).To solve the problem,we propose an energy harvesting based task scheduling and resource management framework to provide robust and low-cost edge computing services for smart grid.First,we formulate an energy consumption minimization problem with regard to task offloading,time switching,and resource allocation for mobile devices,which can be decoupled and transformed into a typical knapsack problem.Then,solutions are derived by two different algorithms.Furthermore,we deploy renewable energy and energy storage units at edge servers to tackle intermittency and instability problems.Finally,we design an energy management algorithm based on sampling average approximation for edge computing servers to derive the optimal charging/discharging strategies,number of energy storage units,and renewable energy utilization.The simulation results show the efficiency and superiority of our proposed framework.展开更多
A nonlinear multi-scale interaction(NMI)model was proposed and developed by the first author for nearly 30 years to represent the evolution of atmospheric blocking.In this review paper,we first review the creation and...A nonlinear multi-scale interaction(NMI)model was proposed and developed by the first author for nearly 30 years to represent the evolution of atmospheric blocking.In this review paper,we first review the creation and development of the NMI model and then emphasize that the NMI model represents a new tool for identifying the basic physics of how climate change influences mid-to-high latitude weather extremes.The building of the NMI model took place over three main periods.In the 1990s,a nonlinear Schr?dinger(NLS)equation model was presented to describe atmospheric blocking as a wave packet;however,it could not depict the lifetime(10-20 days)of atmospheric blocking.In the 2000s,we proposed an NMI model of atmospheric blocking in a uniform basic flow by making a scale-separation assumption and deriving an eddyforced NLS equation.This model succeeded in describing the life cycle of atmospheric blocking.In the 2020s,the NMI model was extended to include the impact of a changing climate mainly by altering the basic zonal winds and the magnitude of the meridional background potential vorticity gradient(PVy).Model results show that when PVy is smaller,blocking has a weaker dispersion and a stronger nonlinearity,so blocking can be more persistent and have a larger zonal scale and weaker eastward movement,thus favoring stronger weather extremes.However,when PVy is much smaller and below a critical threshold under much stronger winter Arctic warming of global warming,atmospheric blocking becomes locally less persistent and shows a much stronger westward movement,which acts to inhibit local cold extremes.Such a case does not happen in summer under global warming because PVy fails to fall below the critical threshold.Thus,our theory indicates that global warming can render summer-blocking anticyclones and mid-to-high latitude heatwaves more persistent,intense,and widespread.展开更多
Deep Learning has been widely used to model soft sensors in modern industrial processes with nonlinear variables and uncertainty.Due to the outstanding ability for high-level feature extraction,stacked autoencoder(SAE...Deep Learning has been widely used to model soft sensors in modern industrial processes with nonlinear variables and uncertainty.Due to the outstanding ability for high-level feature extraction,stacked autoencoder(SAE)has been widely used to improve the model accuracy of soft sensors.However,with the increase of network layers,SAE may encounter serious information loss issues,which affect the modeling performance of soft sensors.Besides,there are typically very few labeled samples in the data set,which brings challenges to traditional neural networks to solve.In this paper,a multi-scale feature fused stacked autoencoder(MFF-SAE)is suggested for feature representation related to hierarchical output,where stacked autoencoder,mutual information(MI)and multi-scale feature fusion(MFF)strategies are integrated.Based on correlation analysis between output and input variables,critical hidden variables are extracted from the original variables in each autoencoder's input layer,which are correspondingly given varying weights.Besides,an integration strategy based on multi-scale feature fusion is adopted to mitigate the impact of information loss with the deepening of the network layers.Then,the MFF-SAE method is designed and stacked to form deep networks.Two practical industrial processes are utilized to evaluate the performance of MFF-SAE.Results from simulations indicate that in comparison to other cutting-edge techniques,the proposed method may considerably enhance the accuracy of soft sensor modeling,where the suggested method reduces the root mean square error(RMSE)by 71.8%,17.1%and 64.7%,15.1%,respectively.展开更多
Large language models(LLMs)have emerged as powerful tools for addressing a wide range of problems,including those in scientific computing,particularly in solving partial differential equations(PDEs).However,different ...Large language models(LLMs)have emerged as powerful tools for addressing a wide range of problems,including those in scientific computing,particularly in solving partial differential equations(PDEs).However,different models exhibit distinct strengths and preferences,resulting in varying levels of performance.In this paper,we compare the capabilities of the most advanced LLMs—DeepSeek,ChatGPT,and Claude—along with their reasoning-optimized versions in addressing computational challenges.Specifically,we evaluate their proficiency in solving traditional numerical problems in scientific computing as well as leveraging scientific machine learning techniques for PDE-based problems.We designed all our experiments so that a nontrivial decision is required,e.g,defining the proper space of input functions for neural operator learning.Our findings show that reasoning and hybrid-reasoning models consistently and significantly outperform non-reasoning ones in solving challenging problems,with ChatGPT o3-mini-high generally offering the fastest reasoning speed.展开更多
Efficient resource provisioning,allocation,and computation offloading are critical to realizing lowlatency,scalable,and energy-efficient applications in cloud,fog,and edge computing.Despite its importance,integrating ...Efficient resource provisioning,allocation,and computation offloading are critical to realizing lowlatency,scalable,and energy-efficient applications in cloud,fog,and edge computing.Despite its importance,integrating Software Defined Networks(SDN)for enhancing resource orchestration,task scheduling,and traffic management remains a relatively underexplored area with significant innovation potential.This paper provides a comprehensive review of existing mechanisms,categorizing resource provisioning approaches into static,dynamic,and user-centric models,while examining applications across domains such as IoT,healthcare,and autonomous systems.The survey highlights challenges such as scalability,interoperability,and security in managing dynamic and heterogeneous infrastructures.This exclusive research evaluates how SDN enables adaptive policy-based handling of distributed resources through advanced orchestration processes.Furthermore,proposes future directions,including AI-driven optimization techniques and hybrid orchestrationmodels.By addressing these emerging opportunities,thiswork serves as a foundational reference for advancing resource management strategies in next-generation cloud,fog,and edge computing ecosystems.This survey concludes that SDN-enabled computing environments find essential guidance in addressing upcoming management opportunities.展开更多
The number of satellites,especially those operating in Low-Earth Orbit(LEO),has been exploding in recent years.Additionally,the burgeoning development of Artificial Intelligence(AI)software and hardware has opened up ...The number of satellites,especially those operating in Low-Earth Orbit(LEO),has been exploding in recent years.Additionally,the burgeoning development of Artificial Intelligence(AI)software and hardware has opened up new industrial opportunities in both air and space,with satellite-powered computing emerging as a new computing paradigm:Orbital Edge Computing(OEC).Compared to terrestrial edge computing,the mobility of LEO satellites and their limited communication,computation,and storage resources pose challenges in designing task-specific scheduling algorithms.Previous survey papers have largely focused on terrestrial edge computing or the integration of space and ground technologies,lacking a comprehensive summary of OEC architecture,algorithms,and case studies.This paper conducts a comprehensive survey and analysis of OEC's system architecture,applications,algorithms,and simulation tools,providing a solid background for researchers in the field.By discussing OEC use cases and the challenges faced,potential research directions for future OEC research are proposed.展开更多
Mental health problems and potential psychological crises affect the healthy growth and learning performance of college students.Effective and suitable prevention of psychological crises among college students is a co...Mental health problems and potential psychological crises affect the healthy growth and learning performance of college students.Effective and suitable prevention of psychological crises among college students is a continuous challenge university managers face.To explore a method of preventing psychological crises among college students,we measured 38661 students by using SCL-90(symptom check list-90)and screened out 5790 students with positive results.Then,we measured 33188 students by using PHQ-9(patient health questionnaire-9)and screened out 603 students with suicidal ideation or behavior;we interviewed 392 students by using GAQ(growth adversity questionnaire).The number of students who had positive results at both phases is 155.As a result,we obtained a data set(N=76)by integrating the students who tested positive on the PHQ-9(i.e.total score≥20)with those who completed the PHQ-9 and GAQ.In addition,we obtained a data set(N=50)by excluding the cases in which the GAQ score is 0.With regard to QCA(qualitative comparative analysis)results,the data set(N=76)exhibits 5 constellations of solutions with a coverage rate greater than 0.7,and the first eight indicators of the PHQ-9 constitute the explanatory variables in the combined solutions.About the data set(N=50),the combined solutions are extremely complicated and the explanatory variables encompass indicators from both the PHQ-9 and GAQ.All these mean that the multi-scale could more comprehensively reflect mental health states of college students,thus enhance the accuracy and effectiveness of the corresponding hierarchical intervention,and finally provide support for preventing psychological crises in universities.展开更多
The emergence of different computing methods such as cloud-,fog-,and edge-based Internet of Things(IoT)systems has provided the opportunity to develop intelligent systems for disease detection.Compared to other machin...The emergence of different computing methods such as cloud-,fog-,and edge-based Internet of Things(IoT)systems has provided the opportunity to develop intelligent systems for disease detection.Compared to other machine learning models,deep learning models have gained more attention from the research community,as they have shown better results with a large volume of data compared to shallow learning.However,no comprehensive survey has been conducted on integrated IoT-and computing-based systems that deploy deep learning for disease detection.This study evaluated different machine learning and deep learning algorithms and their hybrid and optimized algorithms for IoT-based disease detection,using the most recent papers on IoT-based disease detection systems that include computing approaches,such as cloud,edge,and fog.Their analysis focused on an IoT deep learning architecture suitable for disease detection.It also recognizes the different factors that require the attention of researchers to develop better IoT disease detection systems.This study can be helpful to researchers interested in developing better IoT-based disease detection and prediction systems based on deep learning using hybrid algorithms.展开更多
基金supported by the National Natural Science Foundation of China(Grant No.92152301).
文摘Numerical simulation is dominant in solving partial differential equations(PDEs),but balancing fine-grained grids with low computational costs is challenging.Recently,solving PDEs with neural networks(NNs)has gained interest,yet cost-effectiveness and high accuracy remain a challenge.This work introduces a novel paradigm for solving PDEs,called multi-scale neural computing(MSNC),considering spectral bias of NNs and local approximation properties in the finite difference method(FDM).The MSNC decomposes the solution with a NN for efficient capture of global scale and the FDM for detailed description of local scale,aiming to balance costs and accuracy.Demonstrated advantages include higher accuracy(10 times for 1D PDEs,20 times for 2D PDEs)and lower costs(4 times for 1D PDEs,16 times for 2D PDEs)than the standard FDM.The MSNC also exhibits stable convergence and rigorous boundary condition satisfaction,showcasing the potential for hybrid of NN and numerical method.
基金supported by Communication University of China(HG23035)partly supported by the Fundamental Research Funds for the Central Universities(CUC230A013).
文摘With the rapid growth of socialmedia,the spread of fake news has become a growing problem,misleading the public and causing significant harm.As social media content is often composed of both images and text,the use of multimodal approaches for fake news detection has gained significant attention.To solve the problems existing in previous multi-modal fake news detection algorithms,such as insufficient feature extraction and insufficient use of semantic relations between modes,this paper proposes the MFFFND-Co(Multimodal Feature Fusion Fake News Detection with Co-Attention Block)model.First,the model deeply explores the textual content,image content,and frequency domain features.Then,it employs a Co-Attention mechanism for cross-modal fusion.Additionally,a semantic consistency detectionmodule is designed to quantify semantic deviations,thereby enhancing the performance of fake news detection.Experimentally verified on two commonly used datasets,Twitter and Weibo,the model achieved F1 scores of 90.0% and 94.0%,respectively,significantly outperforming the pre-modified MFFFND(Multimodal Feature Fusion Fake News Detection with Attention Block)model and surpassing other baseline models.This improves the accuracy of detecting fake information in artificial intelligence detection and engineering software detection.
文摘Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency.
基金Opening Foundation of Key Laboratory of Explosive Energy Utilization and Control,Anhui Province(BP20240104)Graduate Innovation Program of China University of Mining and Technology(2024WLJCRCZL049)Postgraduate Research&Practice Innovation Program of Jiangsu Province(KYCX24_2701)。
文摘Because of the challenge of compounding lightweight,high-strength Ti/Al alloys due to their considerable disparity in properties,Al 6063 as intermediate layer was proposed to fabricate TC4/Al 6063/Al 7075 three-layer composite plate by explosive welding.The microscopic properties of each bonding interface were elucidated through field emission scanning electron microscope and electron backscattered diffraction(EBSD).A methodology combining finite element method-smoothed particle hydrodynamics(FEM-SPH)and molecular dynamics(MD)was proposed for the analysis of the forming and evolution characteristics of explosive welding interfaces at multi-scale.The results demonstrate that the bonding interface morphologies of TC4/Al 6063 and Al 6063/Al 7075 exhibit a flat and wavy configuration,without discernible defects or cracks.The phenomenon of grain refinement is observed in the vicinity of the two bonding interfaces.Furthermore,the degree of plastic deformation of TC4 and Al 7075 is more pronounced than that of Al 6063 in the intermediate layer.The interface morphology characteristics obtained by FEM-SPH simulation exhibit a high degree of similarity to the experimental results.MD simulations reveal that the diffusion of interfacial elements predominantly occurs during the unloading phase,and the simulated thickness of interfacial diffusion aligns well with experimental outcomes.The introduction of intermediate layer in the explosive welding process can effectively produce high-quality titanium/aluminum alloy composite plates.Furthermore,this approach offers a multi-scale simulation strategy for the study of explosive welding bonding interfaces.
基金supported by the"Science and Technology Development Plan Project of Jilin Province,China"(Grant No.20240101018JJ)the Fundamental Research Funds for the Central Universities(Grant No.2412023YQ004)the National Natural Science Foundation of China(Grant Nos.52072065,52272140,52372137,and U23A20568).
文摘Optoelectronic memristor is generating growing research interest for high efficient computing and sensing-memory applications.In this work,an optoelectronic memristor with Au/a-C:Te/Pt structure is developed.Synaptic functions,i.e.,excita-tory post-synaptic current and pair-pulse facilitation are successfully mimicked with the memristor under electrical and optical stimulations.More importantly,the device exhibited distinguishable response currents by adjusting 4-bit input electrical/opti-cal signals.A multi-mode reservoir computing(RC)system is constructed with the optoelectronic memristors to emulate human tactile-visual fusion recognition and an accuracy of 98.7%is achieved.The optoelectronic memristor provides potential for developing multi-mode RC system.
基金supported by the Natural Science Foundation of China(Grant No.42302170)National Postdoctoral Innovative Talent Support Program(Grant No.BX20220062)+3 种基金CNPC Innovation Found(Grant No.2022DQ02-0104)National Science Foundation of Heilongjiang Province of China(Grant No.YQ2023D001)Postdoctoral Science Foundation of Heilongjiang Province of China(Grant No.LBH-Z22091)the Natural Science Foundation of Shandong Province(Grant No.ZR2022YQ30).
文摘Prediction of production decline and evaluation of the adsorbed/free gas ratio are critical for determining the lifespan and production status of shale gas wells.Traditional production prediction methods have some shortcomings because of the low permeability and tightness of shale,complex gas flow behavior of multi-scale gas transport regions and multiple gas transport mechanism superpositions,and complex and variable production regimes of shale gas wells.Recent research has demonstrated the existence of a multi-stage isotope fractionation phenomenon during shale gas production,with the fractionation characteristics of each stage associated with the pore structure,gas in place(GIP),adsorption/desorption,and gas production process.This study presents a new approach for estimating shale gas well production and evaluating the adsorbed/free gas ratio throughout production using isotope fractionation techniques.A reservoir-scale carbon isotope fractionation(CIF)model applicable to the production process of shale gas wells was developed for the first time in this research.In contrast to the traditional model,this model improves production prediction accuracy by simultaneously fitting the gas production rate and δ^(13)C_(1) data and provides a new evaluation method of the adsorbed/free gas ratio during shale gas production.The results indicate that the diffusion and adsorption/desorption properties of rock,bottom-hole flowing pressure(BHP)of gas well,and multi-scale gas transport regions of the reservoir all affect isotope fractionation,with the diffusion and adsorption/desorption parameters of rock having the greatest effect on isotope fractionation being D∗/D,PL,VL,α,and others in that order.We effectively tested the universality of the four-stage isotope fractionation feature and revealed a unique isotope fractionation mechanism caused by the superimposed coupling of multi-scale gas transport regions during shale gas well production.Finally,we applied the established CIF model to a shale gas well in the Sichuan Basin,China,and calculated the estimated ultimate recovery(EUR)of the well to be 3.33×10^(8) m^(3);the adsorbed gas ratio during shale gas production was 1.65%,10.03%,and 23.44%in the first,fifth,and tenth years,respectively.The findings are significant for understanding the isotope fractionation mechanism during natural gas transport in complex systems and for formulating and optimizing unconventional natural gas development strategies.
基金funded by the Fundamental Research Funds for the Central Universities(J2023-024,J2023-027).
文摘As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the task offloading strategies by interacting with the entities. In actual application scenarios, users of edge computing are always changing dynamically. However, the existing task offloading strategies cannot be applied to such dynamic scenarios. To solve this problem, we propose a novel dynamic task offloading framework for distributed edge computing, leveraging the potential of meta-reinforcement learning (MRL). Our approach formulates a multi-objective optimization problem aimed at minimizing both delay and energy consumption. We model the task offloading strategy using a directed acyclic graph (DAG). Furthermore, we propose a distributed edge computing adaptive task offloading algorithm rooted in MRL. This algorithm integrates multiple Markov decision processes (MDP) with a sequence-to-sequence (seq2seq) network, enabling it to learn and adapt task offloading strategies responsively across diverse network environments. To achieve joint optimization of delay and energy consumption, we incorporate the non-dominated sorting genetic algorithm II (NSGA-II) into our framework. Simulation results demonstrate the superiority of our proposed solution, achieving a 21% reduction in time delay and a 19% decrease in energy consumption compared to alternative task offloading schemes. Moreover, our scheme exhibits remarkable adaptability, responding swiftly to changes in various network environments.
基金supported by the National Natural Science Foundation of China(62272049,62236006,62172045)the Key Projects of Beijing Union University(ZKZD202301).
文摘In recent years,gait-based emotion recognition has been widely applied in the field of computer vision.However,existing gait emotion recognition methods typically rely on complete human skeleton data,and their accuracy significantly declines when the data is occluded.To enhance the accuracy of gait emotion recognition under occlusion,this paper proposes a Multi-scale Suppression Graph ConvolutionalNetwork(MS-GCN).TheMS-GCN consists of three main components:Joint Interpolation Module(JI Moudle),Multi-scale Temporal Convolution Network(MS-TCN),and Suppression Graph Convolutional Network(SGCN).The JI Module completes the spatially occluded skeletal joints using the(K-Nearest Neighbors)KNN interpolation method.The MS-TCN employs convolutional kernels of various sizes to comprehensively capture the emotional information embedded in the gait,compensating for the temporal occlusion of gait information.The SGCN extracts more non-prominent human gait features by suppressing the extraction of key body part features,thereby reducing the negative impact of occlusion on emotion recognition results.The proposed method is evaluated on two comprehensive datasets:Emotion-Gait,containing 4227 real gaits from sources like BML,ICT-Pollick,and ELMD,and 1000 synthetic gaits generated using STEP-Gen technology,and ELMB,consisting of 3924 gaits,with 1835 labeled with emotions such as“Happy,”“Sad,”“Angry,”and“Neutral.”On the standard datasets Emotion-Gait and ELMB,the proposed method achieved accuracies of 0.900 and 0.896,respectively,attaining performance comparable to other state-ofthe-artmethods.Furthermore,on occlusion datasets,the proposedmethod significantly mitigates the performance degradation caused by occlusion compared to other methods,the accuracy is significantly higher than that of other methods.
基金the National Research Foundation(NRF)Singapore mid-sized center grant(NRF-MSG-2023-0002)FrontierCRP grant(NRF-F-CRP-2024-0006)+2 种基金A*STAR Singapore MTC RIE2025 project(M24W1NS005)IAF-PP project(M23M5a0069)Ministry of Education(MOE)Singapore Tier 2 project(MOE-T2EP50220-0014).
文摘The rise of large-scale artificial intelligence(AI)models,such as ChatGPT,Deep-Seek,and autonomous vehicle systems,has significantly advanced the boundaries of AI,enabling highly complex tasks in natural language processing,image recognition,and real-time decisionmaking.However,these models demand immense computational power and are often centralized,relying on cloud-based architectures with inherent limitations in latency,privacy,and energy efficiency.To address these challenges and bring AI closer to real-world applications,such as wearable health monitoring,robotics,and immersive virtual environments,innovative hardware solutions are urgently needed.This work introduces a near-sensor edge computing(NSEC)system,built on a bilayer AlN/Si waveguide platform,to provide real-time,energy-efficient AI capabilities at the edge.Leveraging the electro-optic properties of AlN microring resonators for photonic feature extraction,coupled with Si-based thermo-optic Mach-Zehnder interferometers for neural network computations,the system represents a transformative approach to AI hardware design.Demonstrated through multimodal gesture and gait analysis,the NSEC system achieves high classification accuracies of 96.77%for gestures and 98.31%for gaits,ultra-low latency(<10 ns),and minimal energy consumption(<0.34 pJ).This groundbreaking system bridges the gap between AI models and real-world applications,enabling efficient,privacy-preserving AI solutions for healthcare,robotics,and next-generation human-machine interfaces,marking a pivotal advancement in edge computing and AI deployment.
基金supported by the Natural Science Foundation of Zhejiang Province(Grant No.LQ24F040007)the National Natural Science Foundation of China(Grant No.U22A2075)the Opening Project of State Key Laboratory of Polymer Materials Engineering(Sichuan University)(Grant No.sklpme2024-1-21).
文摘To address the increasing demand for massive data storage and processing,brain-inspired neuromorphic comput-ing systems based on artificial synaptic devices have been actively developed in recent years.Among the various materials inves-tigated for the fabrication of synaptic devices,silicon carbide(SiC)has emerged as a preferred choices due to its high electron mobility,superior thermal conductivity,and excellent thermal stability,which exhibits promising potential for neuromorphic applications in harsh environments.In this review,the recent progress in SiC-based synaptic devices is summarized.Firstly,an in-depth discussion is conducted regarding the categories,working mechanisms,and structural designs of these devices.Subse-quently,several application scenarios for SiC-based synaptic devices are presented.Finally,a few perspectives and directions for their future development are outlined.
基金Supported by the Henan Province Key Research and Development Project(231111211300)the Central Government of Henan Province Guides Local Science and Technology Development Funds(Z20231811005)+2 种基金Henan Province Key Research and Development Project(231111110100)Henan Provincial Outstanding Foreign Scientist Studio(GZS2024006)Henan Provincial Joint Fund for Scientific and Technological Research and Development Plan(Application and Overcoming Technical Barriers)(242103810028)。
文摘The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method for infrared and visible image fusion is proposed.The encoder designed according to the optimization objective consists of a base encoder and a detail encoder,which is used to extract low-frequency and high-frequency information from the image.This extraction may lead to some information not being captured,so a compensation encoder is proposed to supplement the missing information.Multi-scale decomposition is also employed to extract image features more comprehensively.The decoder combines low-frequency,high-frequency and supplementary information to obtain multi-scale features.Subsequently,the attention strategy and fusion module are introduced to perform multi-scale fusion for image reconstruction.Experimental results on three datasets show that the fused images generated by this network effectively retain salient targets while being more consistent with human visual perception.
文摘The rapid advent in artificial intelligence and big data has revolutionized the dynamic requirement in the demands of the computing resource for executing specific tasks in the cloud environment.The process of achieving autonomic resource management is identified to be a herculean task due to its huge distributed and heterogeneous environment.Moreover,the cloud network needs to provide autonomic resource management and deliver potential services to the clients by complying with the requirements of Quality-of-Service(QoS)without impacting the Service Level Agreements(SLAs).However,the existing autonomic cloud resource managing frameworks are not capable in handling the resources of the cloud with its dynamic requirements.In this paper,Coot Bird Behavior Model-based Workload Aware Autonomic Resource Management Scheme(CBBM-WARMS)is proposed for handling the dynamic requirements of cloud resources through the estimation of workload that need to be policed by the cloud environment.This CBBM-WARMS initially adopted the algorithm of adaptive density peak clustering for workloads clustering of the cloud.Then,it utilized the fuzzy logic during the process of workload scheduling for achieving the determining the availability of cloud resources.It further used CBBM for potential Virtual Machine(VM)deployment that attributes towards the provision of optimal resources.It is proposed with the capability of achieving optimal QoS with minimized time,energy consumption,SLA cost and SLA violation.The experimental validation of the proposed CBBMWARMS confirms minimized SLA cost of 19.21%and reduced SLA violation rate of 18.74%,better than the compared autonomic cloud resource managing frameworks.
基金supported in part by the National Natural Science Foundation of China under Grant No.61473066in part by the Natural Science Foundation of Hebei Province under Grant No.F2021501020+2 种基金in part by the S&T Program of Qinhuangdao under Grant No.202401A195in part by the Science Research Project of Hebei Education Department under Grant No.QN2025008in part by the Innovation Capability Improvement Plan Project of Hebei Province under Grant No.22567637H
文摘Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power stations).To solve the problem,we propose an energy harvesting based task scheduling and resource management framework to provide robust and low-cost edge computing services for smart grid.First,we formulate an energy consumption minimization problem with regard to task offloading,time switching,and resource allocation for mobile devices,which can be decoupled and transformed into a typical knapsack problem.Then,solutions are derived by two different algorithms.Furthermore,we deploy renewable energy and energy storage units at edge servers to tackle intermittency and instability problems.Finally,we design an energy management algorithm based on sampling average approximation for edge computing servers to derive the optimal charging/discharging strategies,number of energy storage units,and renewable energy utilization.The simulation results show the efficiency and superiority of our proposed framework.
基金supported by the National Natural Science Foundation of China(Grant Nos.42150204 and 2288101)supported by the China National Postdoctoral Program for Innovative Talents(BX20230045)the China Postdoctoral Science Foundation(2023M730279)。
文摘A nonlinear multi-scale interaction(NMI)model was proposed and developed by the first author for nearly 30 years to represent the evolution of atmospheric blocking.In this review paper,we first review the creation and development of the NMI model and then emphasize that the NMI model represents a new tool for identifying the basic physics of how climate change influences mid-to-high latitude weather extremes.The building of the NMI model took place over three main periods.In the 1990s,a nonlinear Schr?dinger(NLS)equation model was presented to describe atmospheric blocking as a wave packet;however,it could not depict the lifetime(10-20 days)of atmospheric blocking.In the 2000s,we proposed an NMI model of atmospheric blocking in a uniform basic flow by making a scale-separation assumption and deriving an eddyforced NLS equation.This model succeeded in describing the life cycle of atmospheric blocking.In the 2020s,the NMI model was extended to include the impact of a changing climate mainly by altering the basic zonal winds and the magnitude of the meridional background potential vorticity gradient(PVy).Model results show that when PVy is smaller,blocking has a weaker dispersion and a stronger nonlinearity,so blocking can be more persistent and have a larger zonal scale and weaker eastward movement,thus favoring stronger weather extremes.However,when PVy is much smaller and below a critical threshold under much stronger winter Arctic warming of global warming,atmospheric blocking becomes locally less persistent and shows a much stronger westward movement,which acts to inhibit local cold extremes.Such a case does not happen in summer under global warming because PVy fails to fall below the critical threshold.Thus,our theory indicates that global warming can render summer-blocking anticyclones and mid-to-high latitude heatwaves more persistent,intense,and widespread.
基金supported by the National Key Research and Development Program of China(2023YFB3307800)National Natural Science Foundation of China(62394343,62373155)+2 种基金Major Science and Technology Project of Xinjiang(No.2022A01006-4)State Key Laboratory of Industrial Control Technology,China(Grant No.ICT2024A26)Fundamental Research Funds for the Central Universities.
文摘Deep Learning has been widely used to model soft sensors in modern industrial processes with nonlinear variables and uncertainty.Due to the outstanding ability for high-level feature extraction,stacked autoencoder(SAE)has been widely used to improve the model accuracy of soft sensors.However,with the increase of network layers,SAE may encounter serious information loss issues,which affect the modeling performance of soft sensors.Besides,there are typically very few labeled samples in the data set,which brings challenges to traditional neural networks to solve.In this paper,a multi-scale feature fused stacked autoencoder(MFF-SAE)is suggested for feature representation related to hierarchical output,where stacked autoencoder,mutual information(MI)and multi-scale feature fusion(MFF)strategies are integrated.Based on correlation analysis between output and input variables,critical hidden variables are extracted from the original variables in each autoencoder's input layer,which are correspondingly given varying weights.Besides,an integration strategy based on multi-scale feature fusion is adopted to mitigate the impact of information loss with the deepening of the network layers.Then,the MFF-SAE method is designed and stacked to form deep networks.Two practical industrial processes are utilized to evaluate the performance of MFF-SAE.Results from simulations indicate that in comparison to other cutting-edge techniques,the proposed method may considerably enhance the accuracy of soft sensor modeling,where the suggested method reduces the root mean square error(RMSE)by 71.8%,17.1%and 64.7%,15.1%,respectively.
基金supported by the ONR Vannevar Bush Faculty Fellowship(Grant No.N00014-22-1-2795).
文摘Large language models(LLMs)have emerged as powerful tools for addressing a wide range of problems,including those in scientific computing,particularly in solving partial differential equations(PDEs).However,different models exhibit distinct strengths and preferences,resulting in varying levels of performance.In this paper,we compare the capabilities of the most advanced LLMs—DeepSeek,ChatGPT,and Claude—along with their reasoning-optimized versions in addressing computational challenges.Specifically,we evaluate their proficiency in solving traditional numerical problems in scientific computing as well as leveraging scientific machine learning techniques for PDE-based problems.We designed all our experiments so that a nontrivial decision is required,e.g,defining the proper space of input functions for neural operator learning.Our findings show that reasoning and hybrid-reasoning models consistently and significantly outperform non-reasoning ones in solving challenging problems,with ChatGPT o3-mini-high generally offering the fastest reasoning speed.
文摘Efficient resource provisioning,allocation,and computation offloading are critical to realizing lowlatency,scalable,and energy-efficient applications in cloud,fog,and edge computing.Despite its importance,integrating Software Defined Networks(SDN)for enhancing resource orchestration,task scheduling,and traffic management remains a relatively underexplored area with significant innovation potential.This paper provides a comprehensive review of existing mechanisms,categorizing resource provisioning approaches into static,dynamic,and user-centric models,while examining applications across domains such as IoT,healthcare,and autonomous systems.The survey highlights challenges such as scalability,interoperability,and security in managing dynamic and heterogeneous infrastructures.This exclusive research evaluates how SDN enables adaptive policy-based handling of distributed resources through advanced orchestration processes.Furthermore,proposes future directions,including AI-driven optimization techniques and hybrid orchestrationmodels.By addressing these emerging opportunities,thiswork serves as a foundational reference for advancing resource management strategies in next-generation cloud,fog,and edge computing ecosystems.This survey concludes that SDN-enabled computing environments find essential guidance in addressing upcoming management opportunities.
基金funded by the Hong Kong-Macao-Taiwan Science and Technology Cooperation Project of the Science and Technology Innovation Action Plan in Shanghai,China(23510760200)the Oriental Talent Youth Program of Shanghai,China(No.Y3DFRCZL01)+1 种基金the Outstanding Program of the Youth Innovation Promotion Association of the Chinese Academy of Sciences(No.Y2023080)the Strategic Priority Research Program of the Chinese Academy of Sciences Category A(No.XDA0360404).
文摘The number of satellites,especially those operating in Low-Earth Orbit(LEO),has been exploding in recent years.Additionally,the burgeoning development of Artificial Intelligence(AI)software and hardware has opened up new industrial opportunities in both air and space,with satellite-powered computing emerging as a new computing paradigm:Orbital Edge Computing(OEC).Compared to terrestrial edge computing,the mobility of LEO satellites and their limited communication,computation,and storage resources pose challenges in designing task-specific scheduling algorithms.Previous survey papers have largely focused on terrestrial edge computing or the integration of space and ground technologies,lacking a comprehensive summary of OEC architecture,algorithms,and case studies.This paper conducts a comprehensive survey and analysis of OEC's system architecture,applications,algorithms,and simulation tools,providing a solid background for researchers in the field.By discussing OEC use cases and the challenges faced,potential research directions for future OEC research are proposed.
文摘Mental health problems and potential psychological crises affect the healthy growth and learning performance of college students.Effective and suitable prevention of psychological crises among college students is a continuous challenge university managers face.To explore a method of preventing psychological crises among college students,we measured 38661 students by using SCL-90(symptom check list-90)and screened out 5790 students with positive results.Then,we measured 33188 students by using PHQ-9(patient health questionnaire-9)and screened out 603 students with suicidal ideation or behavior;we interviewed 392 students by using GAQ(growth adversity questionnaire).The number of students who had positive results at both phases is 155.As a result,we obtained a data set(N=76)by integrating the students who tested positive on the PHQ-9(i.e.total score≥20)with those who completed the PHQ-9 and GAQ.In addition,we obtained a data set(N=50)by excluding the cases in which the GAQ score is 0.With regard to QCA(qualitative comparative analysis)results,the data set(N=76)exhibits 5 constellations of solutions with a coverage rate greater than 0.7,and the first eight indicators of the PHQ-9 constitute the explanatory variables in the combined solutions.About the data set(N=50),the combined solutions are extremely complicated and the explanatory variables encompass indicators from both the PHQ-9 and GAQ.All these mean that the multi-scale could more comprehensively reflect mental health states of college students,thus enhance the accuracy and effectiveness of the corresponding hierarchical intervention,and finally provide support for preventing psychological crises in universities.
文摘The emergence of different computing methods such as cloud-,fog-,and edge-based Internet of Things(IoT)systems has provided the opportunity to develop intelligent systems for disease detection.Compared to other machine learning models,deep learning models have gained more attention from the research community,as they have shown better results with a large volume of data compared to shallow learning.However,no comprehensive survey has been conducted on integrated IoT-and computing-based systems that deploy deep learning for disease detection.This study evaluated different machine learning and deep learning algorithms and their hybrid and optimized algorithms for IoT-based disease detection,using the most recent papers on IoT-based disease detection systems that include computing approaches,such as cloud,edge,and fog.Their analysis focused on an IoT deep learning architecture suitable for disease detection.It also recognizes the different factors that require the attention of researchers to develop better IoT disease detection systems.This study can be helpful to researchers interested in developing better IoT-based disease detection and prediction systems based on deep learning using hybrid algorithms.