The rise of time-sensitive applications with broad geographical scope drives the development of time-sensitive networking(TSN)from intra-domain to inter-domain to ensure overall end-to-end connectivity requirements in...The rise of time-sensitive applications with broad geographical scope drives the development of time-sensitive networking(TSN)from intra-domain to inter-domain to ensure overall end-to-end connectivity requirements in heterogeneous deployments.When multiple TSN networks interconnect over non-TSN networks,all devices in the network need to be syn-chronized by sharing a uniform time reference.How-ever,most non-TSN networks are best-effort.Path delay asymmetry and random noise accumulation can introduce unpredictable time errors during end-to-end time synchronization.These factors can degrade syn-chronization performance.Therefore,cross-domain time synchronization becomes a challenging issue for multiple TSN networks interconnected by non-TSN networks.This paper presents a cross-domain time synchronization scheme that follows the software-defined TSN(SD-TSN)paradigm.It utilizes a com-bined control plane constructed by a coordinate con-troller and a domain controller for centralized control and management of cross-domain time synchroniza-tion.The general operation flow of the cross-domain time synchronization process is designed.The mecha-nism of cross-domain time synchronization is revealed by introducing a synchronization model and an error compensation method.A TSN cross-domain proto-type testbed is constructed for verification.Results show that the scheme can achieve end-to-end high-precision time synchronization with accuracy and sta-bility.展开更多
Accurate time synchronization is fundamental to the correct and efficient operation of Wireless Sensor Networks(WSNs),especially in security-critical,time-sensitive applications.However,most existing protocols degrade...Accurate time synchronization is fundamental to the correct and efficient operation of Wireless Sensor Networks(WSNs),especially in security-critical,time-sensitive applications.However,most existing protocols degrade substantially under malicious interference.We introduce iSTSP,an Intelligent and Secure Time Synchronization Protocol that implements a four-stage defense pipeline to ensure robust,precise synchronization even in hostile environments:(1)trust preprocessing that filters node participation using behavioral trust scoring;(2)anomaly isolation employing a lightweight autoencoder to detect and excise malicious nodes in real time;(3)reliability-weighted consensus that prioritizes high-trust nodes during time aggregation;and(4)convergence-optimized synchronization that dynamically adjusts parameters using theoretical stability bounds.We provide rigorous convergence analysis including a closed-form expression for convergence time,and validate the protocol through both simulations and realworld experiments on a controlled 16-node testbed.Under Sybil attacks with five malicious nodes within this testbed,iSTSP maintains synchronization error increases under 12%and achieves a rapid convergence.Compared to state-ofthe-art protocols like TPSN,SE-FTSP,and MMAR-CTS,iSTSP offers 60%faster detection,broader threat coverage,and more than 7 times lower synchronization error,with a modest 9.3%energy overhead over 8 h.We argue this is an acceptable trade-off for mission-critical deployments requiring guaranteed security.These findings demonstrate iSTSP’s potential as a reliable solution for secure WSN synchronization and motivate future work on large-scale IoT deployments and integration with energy-efficient communication protocols.展开更多
Time-Sensitive Network(TSN)with deterministic transmission capability is increasingly used in many emerging fields.It mainly guarantees the Quality of Service(QoS)of applications with strict requirements on time and s...Time-Sensitive Network(TSN)with deterministic transmission capability is increasingly used in many emerging fields.It mainly guarantees the Quality of Service(QoS)of applications with strict requirements on time and security.One of the core features of TSN is traffic scheduling with bounded low delay in the network.However,traffic scheduling schemes in TSN are usually synthesized offline and lack dynamism.To implement incremental scheduling of newly arrived traffic in TSN,we propose a Dynamic Response Incremental Scheduling(DR-IS)method for time-sensitive traffic and deploy it on a software-defined time-sensitive network architecture.Under the premise of meeting the traffic scheduling requirements,we adopt two modes,traffic shift and traffic exchange,to dynamically adjust the time slot injection position of the traffic in the original scheme,and determine the sending offset time of the new timesensitive traffic to minimize the global traffic transmission jitter.The evaluation results show that DRIS method can effectively control the large increase of traffic transmission jitter in incremental scheduling without affecting the transmission delay,thus realizing the dynamic incremental scheduling of time-sensitive traffic in TSN.展开更多
As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the...As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the task offloading strategies by interacting with the entities. In actual application scenarios, users of edge computing are always changing dynamically. However, the existing task offloading strategies cannot be applied to such dynamic scenarios. To solve this problem, we propose a novel dynamic task offloading framework for distributed edge computing, leveraging the potential of meta-reinforcement learning (MRL). Our approach formulates a multi-objective optimization problem aimed at minimizing both delay and energy consumption. We model the task offloading strategy using a directed acyclic graph (DAG). Furthermore, we propose a distributed edge computing adaptive task offloading algorithm rooted in MRL. This algorithm integrates multiple Markov decision processes (MDP) with a sequence-to-sequence (seq2seq) network, enabling it to learn and adapt task offloading strategies responsively across diverse network environments. To achieve joint optimization of delay and energy consumption, we incorporate the non-dominated sorting genetic algorithm II (NSGA-II) into our framework. Simulation results demonstrate the superiority of our proposed solution, achieving a 21% reduction in time delay and a 19% decrease in energy consumption compared to alternative task offloading schemes. Moreover, our scheme exhibits remarkable adaptability, responding swiftly to changes in various network environments.展开更多
Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power sta...Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power stations).To solve the problem,we propose an energy harvesting based task scheduling and resource management framework to provide robust and low-cost edge computing services for smart grid.First,we formulate an energy consumption minimization problem with regard to task offloading,time switching,and resource allocation for mobile devices,which can be decoupled and transformed into a typical knapsack problem.Then,solutions are derived by two different algorithms.Furthermore,we deploy renewable energy and energy storage units at edge servers to tackle intermittency and instability problems.Finally,we design an energy management algorithm based on sampling average approximation for edge computing servers to derive the optimal charging/discharging strategies,number of energy storage units,and renewable energy utilization.The simulation results show the efficiency and superiority of our proposed framework.展开更多
With the rapid expansion of social media,analyzing emotions and their causes in texts has gained significant importance.Emotion-cause pair extraction enables the identification of causal relationships between emotions...With the rapid expansion of social media,analyzing emotions and their causes in texts has gained significant importance.Emotion-cause pair extraction enables the identification of causal relationships between emotions and their triggers within a text,facilitating a deeper understanding of expressed sentiments and their underlying reasons.This comprehension is crucial for making informed strategic decisions in various business and societal contexts.However,recent research approaches employing multi-task learning frameworks for modeling often face challenges such as the inability to simultaneouslymodel extracted features and their interactions,or inconsistencies in label prediction between emotion-cause pair extraction and independent assistant tasks like emotion and cause extraction.To address these issues,this study proposes an emotion-cause pair extraction methodology that incorporates joint feature encoding and task alignment mechanisms.The model consists of two primary components:First,joint feature encoding simultaneously generates features for emotion-cause pairs and clauses,enhancing feature interactions between emotion clauses,cause clauses,and emotion-cause pairs.Second,the task alignment technique is applied to reduce the labeling distance between emotion-cause pair extraction and the two assistant tasks,capturing deep semantic information interactions among tasks.The proposed method is evaluated on a Chinese benchmark corpus using 10-fold cross-validation,assessing key performance metrics such as precision,recall,and F1 score.Experimental results demonstrate that the model achieves an F1 score of 76.05%,surpassing the state-of-the-art by 1.03%.The proposed model exhibits significant improvements in emotion-cause pair extraction(ECPE)and cause extraction(CE)compared to existing methods,validating its effectiveness.This research introduces a novel approach based on joint feature encoding and task alignment mechanisms,contributing to advancements in emotion-cause pair extraction.However,the study’s limitation lies in the data sources,potentially restricting the generalizability of the findings.展开更多
Blockchain technology,based on decentralized data storage and distributed consensus design,has become a promising solution to address data security risks and provide privacy protection in the Internet-of-Things(IoT)du...Blockchain technology,based on decentralized data storage and distributed consensus design,has become a promising solution to address data security risks and provide privacy protection in the Internet-of-Things(IoT)due to its tamper-proof and non-repudiation features.Although blockchain typically does not require the endorsement of third-party trust organizations,it mostly needs to perform necessary mathematical calculations to prevent malicious attacks,which results in stricter requirements for computation resources on the participating devices.By offloading the computation tasks required to support blockchain consensus to edge service nodes or the cloud,while providing data privacy protection for IoT applications,it can effectively address the limitations of computation and energy resources in IoT devices.However,how to make reasonable offloading decisions for IoT devices remains an open issue.Due to the excellent self-learning ability of Reinforcement Learning(RL),this paper proposes a RL enabled Swarm Intelligence Optimization Algorithm(RLSIOA)that aims to improve the quality of initial solutions and achieve efficient optimization of computation task offloading decisions.The algorithm considers various factors that may affect the revenue obtained by IoT devices executing consensus algorithms(e.g.,Proof-of-Work),it optimizes the proportion of sub-tasks to be offloaded and the scale of computing resources to be rented from the edge and cloud to maximize the revenue of devices.Experimental results show that RLSIOA can obtain higher-quality offloading decision-making schemes at lower latency costs compared to representative benchmark algorithms.展开更多
Fog computing is a key enabling technology of 6G systems as it provides quick and reliable computing,and data storage services which are required for several 6G applications.Artificial Intelligence(AI)algorithms will ...Fog computing is a key enabling technology of 6G systems as it provides quick and reliable computing,and data storage services which are required for several 6G applications.Artificial Intelligence(AI)algorithms will be an integral part of 6G systems and efficient task offloading techniques using fog computing will improve their performance and reliability.In this paper,the focus is on the scenario of Partial Offloading of a Task to Multiple Helpers(POMH)in which larger tasks are divided into smaller subtasks and processed in parallel,hence expediting task completion.However,using POMH presents challenges such as breaking tasks into subtasks and scaling these subtasks based on many interdependent factors to ensure that all subtasks of a task finish simultaneously,preventing resource wastage.Additionally,applying matching theory to POMH scenarios results in dynamic preference profiles of helping devices due to changing subtask sizes,resulting in a difficult-to-solve,externalities problem.This paper introduces a novel many-to-one matching-based algorithm,designed to address the externalities problem and optimize resource allocation within POMH scenarios.Additionally,we propose a new time-efficient preference profiling technique that further enhances time optimization in POMH scenarios.The performance of the proposed technique is thoroughly evaluated in comparison to alternate baseline schemes,revealing many advantages of the proposed approach.The simulation findings indisputably show that the proposed matching-based offloading technique outperforms existing methodologies in the literature,yielding a remarkable 52 reduction in task latency,particularly under high workloads.展开更多
With the development of vehicle networks and the construction of roadside units,Vehicular Ad Hoc Networks(VANETs)are increasingly promoting cooperative computing patterns among vehicles.Vehicular edge computing(VEC)of...With the development of vehicle networks and the construction of roadside units,Vehicular Ad Hoc Networks(VANETs)are increasingly promoting cooperative computing patterns among vehicles.Vehicular edge computing(VEC)offers an effective solution to mitigate resource constraints by enabling task offloading to edge cloud infrastructure,thereby reducing the computational burden on connected vehicles.However,this sharing-based and distributed computing paradigm necessitates ensuring the credibility and reliability of various computation nodes.Existing vehicular edge computing platforms have not adequately considered themisbehavior of vehicles.We propose a practical task offloading algorithm based on reputation assessment to address the task offloading problem in vehicular edge computing under an unreliable environment.This approach integrates deep reinforcement learning and reputation management to address task offloading challenges.Simulation experiments conducted using Veins demonstrate the feasibility and effectiveness of the proposed method.展开更多
Drawing upon self-determination theory,this study examines the effects of vicarious abusive supervision on third-party’s self-efficacy and task performance within organizational contexts.Data were collected via surve...Drawing upon self-determination theory,this study examines the effects of vicarious abusive supervision on third-party’s self-efficacy and task performance within organizational contexts.Data were collected via surveys from 337 employees across diverse organizations.The results indicate that vicarious abusive supervision significantly undermines both self-efficacy and task performance among employees who are indirectly exposed to such behavior but not directly targeted.Furthermore,self-efficacy serves as a mediator between vicarious abusive supervision and task performance;however,this mediating effect is attenuated for employees with a high promotion focus.These findings provide valuable theoretical and practical insights,particularly in the domain of organizational behavior,by emphasizing the critical role of promotion focus in mitigating the negative effects of vicarious abusive supervision.This research contributes to the organizational behavior literature by shifting the focus from the traditional supervisor-subordinate dynamic to a third-party perspective,thereby enriching our understanding of how vicarious abusive supervision impacts employees within organizational settings.The study underscores the importance of self-efficacy and promotion focus as key factors in unethical leadership contexts.展开更多
Porters play a crucial role in hospitals because they ensure the efficient transportation of patients,medical equipment,and vital documents.Despite its importance,there is a lack of research addressing the prediction ...Porters play a crucial role in hospitals because they ensure the efficient transportation of patients,medical equipment,and vital documents.Despite its importance,there is a lack of research addressing the prediction of completion times for porter tasks.To address this gap,we utilized real-world porter delivery data from Taiwan University Hospital,China,Yunlin Branch,Taiwan Region of China.We first identified key features that can influence the duration of porter tasks.We then employed three widely-used machine learning algorithms:decision tree,random forest,and gradient boosting.To leverage the strengths of each algorithm,we finally adopted an ensemble modeling approach that aggregates their individual predictions.Our experimental results show that the proposed ensemble model can achieve a mean absolute error of 3 min in predicting task response time and 4.42 min in task completion time.The prediction error is around 50%lower compared to using only the historical average.These results demonstrate that our method significantly improves the accuracy of porter task time prediction,supporting better resource planning and patient care.It helps ward staff streamline workflows by reducing delays,enables porter managers to allocate resources more effectively,and shortens patient waiting times,contributing to a better care experience.展开更多
Under the background of training practical compound talents in software engineering,this paper analyzes the problems existing in the current teaching of software engineering courses represented by software project man...Under the background of training practical compound talents in software engineering,this paper analyzes the problems existing in the current teaching of software engineering courses represented by software project management,puts forward the team task mechanism of software engineering courses with AI empowerment and cooperation and competition,develops a unified project management platform to support the implementation of team tasks,and proves the effectiveness of the scheme through the results obtained.展开更多
The construction of ideological-political education within foreign language courses requires an integrated approach that encompasses value shaping,knowledge transfer,and competence cultivation.A critical challenge in ...The construction of ideological-political education within foreign language courses requires an integrated approach that encompasses value shaping,knowledge transfer,and competence cultivation.A critical challenge in this domain is the effective design and implementation of tasks that instill values,while also synergizing with acquiring knowledge and enhancing competencies.This paper delves into the philosophical underpinnings and operational principles of value-shaping task design and its practical application within the context of foreign language teaching.Utilizing Contemporary College English(Integrated Coursebook 3)as a case study,the paper explores value shaping-based task design in ideological-political education of foreign language courses,with the aim of providing references for the construction of ideological-political education in foreign language teaching.展开更多
This paper considers the swarm vigilance problem for multi-agent systems(MAS),where multiple agents are deployed within a rectangular region for perception-based vigilance.There are two main challenges,namely the task...This paper considers the swarm vigilance problem for multi-agent systems(MAS),where multiple agents are deployed within a rectangular region for perception-based vigilance.There are two main challenges,namely the task allocation for vigilance roles and the coverage planning of the perception ranges.Firstly,vigilance behavioral patterns and processes in animal populations within natural habitats are investigated.Inspired by these biological vigilance behaviors,an efficient vigilance task allocation model for MAS is proposed.Secondly,the subsequent optimization of task layouts can achieve efficient surveillance coverage with fewer agents,minimizing resource consumption.Thirdly,an improved particle swarm optimization(IPSO)algorithm is proposed,which incorporates fitness-driven adaptive inertia weight dynamics.According to simulation analysis and comparative studies,optimal parameter configurations for genetic algorithm(GA)and IPSO are determined.Finally,the results indicate the proposed IPSO's superior performance to both GA and standard particle swarm optimization(PSO)in vigilance task allocation optimization,with satisfying advantages in computational efficiency and solution quality.展开更多
Cybertwin-enabled 6th Generation(6G)network is envisioned to support artificial intelligence-native management to meet changing demands of 6G applications.Multi-Agent Deep Reinforcement Learning(MADRL)technologies dri...Cybertwin-enabled 6th Generation(6G)network is envisioned to support artificial intelligence-native management to meet changing demands of 6G applications.Multi-Agent Deep Reinforcement Learning(MADRL)technologies driven by Cybertwins have been proposed for adaptive task offloading strategies.However,the existence of random transmission delay between Cybertwin-driven agents and underlying networks is not considered in related works,which destroys the standard Markov property and increases the decision reaction time to reduce the task offloading strategy performance.In order to address this problem,we propose a pipelining task offloading method to lower the decision reaction time and model it as a delay-aware Markov Decision Process(MDP).Then,we design a delay-aware MADRL algorithm to minimize the weighted sum of task execution latency and energy consumption.Firstly,the state space is augmented using the lastly-received state and historical actions to rebuild the Markov property.Secondly,Gate Transformer-XL is introduced to capture historical actions'importance and maintain the consistent input dimension dynamically changed due to random transmission delays.Thirdly,a sampling method and a new loss function with the difference between the current and target state value and the difference between real state-action value and augmented state-action value are designed to obtain state transition trajectories close to the real ones.Numerical results demonstrate that the proposed methods are effective in reducing reaction time and improving the task offloading performance in the random-delay Cybertwin-enabled 6G networks.展开更多
Due to the intense data flow in expanding Internet of Things(IoT)applications,a heavy processing cost and workload on the fog-cloud side become inevitable.One of the most critical challenges is optimal task scheduling...Due to the intense data flow in expanding Internet of Things(IoT)applications,a heavy processing cost and workload on the fog-cloud side become inevitable.One of the most critical challenges is optimal task scheduling.Since this is an NP-hard problem type,a metaheuristic approach can be a good option.This study introduces a novel enhancement to the Artificial Rabbits Optimization(ARO)algorithm by integrating Chaotic maps and Levy flight strategies(CLARO).This dual approach addresses the limitations of standard ARO in terms of population diversity and convergence speed.It is designed for task scheduling in fog-cloud environments,optimizing energy consumption,makespan,and execution time simultaneously three critical parameters often treated individually in prior works.Unlike conventional single-objective methods,the proposed approach incorporates a multi-objective fitness function that dynamically adjusts the weight of each parameter,resulting in better resource allocation and load balancing.In analysis,a real-world dataset,the Open-source Google Cloud Jobs Dataset(GoCJ_Dataset),is used for performance measurement,and analyses are performed on three considered parameters.Comparisons are applied with well-known algorithms:GWO,SCSO,PSO,WOA,and ARO to indicate the reliability of the proposed method.In this regard,performance evaluation is performed by assigning these tasks to Virtual Machines(VMs)in the resource pool.Simulations are performed on 90 base cases and 30 scenarios for each evaluation parameter.The results indicated that the proposed algorithm achieved the best makespan performance in 80% of cases,ranked first in execution time in 61%of cases,and performed best in the final parameter in 69% of cases.In addition,according to the obtained results based on the defined fitness function,the proposed method(CLARO)is 2.52%better than ARO,3.95%better than SCSO,5.06%better than GWO,8.15%better than PSO,and 9.41%better than WOA.展开更多
BACKGROUND The neural mechanisms underlying aggressive behavior in schizophrenia(SCZ)remain poorly understood.To date,no studies have reported on the event-related potential(ERP)characteristics of aggression in SCZ us...BACKGROUND The neural mechanisms underlying aggressive behavior in schizophrenia(SCZ)remain poorly understood.To date,no studies have reported on the event-related potential(ERP)characteristics of aggression in SCZ using the competitive reaction time task(CRTT).Further investigation into the ERP correlates of aggression in SCZ would provide valuable insights into the neural processes involved.AIM To explore the neural mechanism of aggressive behavior in SCZ.METHODS Participants of this study included 40 SCZ patients and 42 healthy controls(HCs).The Reactive Proactive Aggression Questionnaire was used to assess trait of aggression.The Barratt Impulsiveness Scale 11 was used to measure impulsiveness.The Positive and Negative Symptom Scale(PANSS)was used to evaluate psychopathological features and disease severity.All participants were measured with ERP while performing the CRTT.Data of behavior,ERP components(P2,N2,and P3),and feedback-related negativity(FRN)were analyzed.RESULTS Analysis of the behavioral data revealed that compared with HCs,SCZ patients exhibited higher punishment choices.Analysis of ERP components showed that compared with HCs,SCZ patients exhibited higher N2 amplitudes and P2 amplitudes during the decision phase of the CRTT;however,SCZ patients exhibited lower FRN amplitudes and lower P3 amplitudes during the outcome phase of the CRTT.The N2 amplitudes evoked by highintensity provocation were positively related to PANSS-P scores.And the P3 amplitudes evoked in the winning trials were negatively correlated with the PANSS-G scores.CONCLUSION SCZ patients exhibit abnormal ERP characteristics evoked by the CRTT,which suggests the neural correlates of aggressive behavior in SCZ.展开更多
Edge computing has transformed smart grids by lowering latency,reducing network congestion,and enabling real-time decision-making.Nevertheless,devising an optimal task-offloading strategy remains challenging,as it mus...Edge computing has transformed smart grids by lowering latency,reducing network congestion,and enabling real-time decision-making.Nevertheless,devising an optimal task-offloading strategy remains challenging,as it must jointly minimise energy consumption and response time under fluctuating workloads and volatile network conditions.We cast the offloading problem as aMarkov Decision Process(MDP)and solve it with Deep Reinforcement Learning(DRL).Specifically,we present a three-tier architecture—end devices,edge nodes,and a cloud server—and enhance Proximal Policy Optimization(PPO)to learn adaptive,energy-aware policies.A Convolutional Neural Network(CNN)extracts high-level features from system states,enabling the agent to respond continually to changing conditions.Extensive simulations show that the proposed method reduces task latency and energy consumption far more than several baseline algorithms,thereby improving overall system performance.These results demonstrate the effectiveness and robustness of the framework for real-time task offloading in dynamic smart-grid environments.展开更多
The widespread adoption of cloud computing has underscored the critical importance of efficient resource allocation and management, particularly in task scheduling, which involves assigning tasks to computing resource...The widespread adoption of cloud computing has underscored the critical importance of efficient resource allocation and management, particularly in task scheduling, which involves assigning tasks to computing resources for optimized resource utilization. Several meta-heuristic algorithms have shown effectiveness in task scheduling, among which the relatively recent Willow Catkin Optimization (WCO) algorithm has demonstrated potential, albeit with apparent needs for enhanced global search capability and convergence speed. To address these limitations of WCO in cloud computing task scheduling, this paper introduces an improved version termed the Advanced Willow Catkin Optimization (AWCO) algorithm. AWCO enhances the algorithm’s performance by augmenting its global search capability through a quasi-opposition-based learning strategy and accelerating its convergence speed via sinusoidal mapping. A comprehensive evaluation utilizing the CEC2014 benchmark suite, comprising 30 test functions, demonstrates that AWCO achieves superior optimization outcomes, surpassing conventional WCO and a range of established meta-heuristics. The proposed algorithm also considers trade-offs among the cost, makespan, and load balancing objectives. Experimental results of AWCO are compared with those obtained using the other meta-heuristics, illustrating that the proposed algorithm provides superior performance in task scheduling. The method offers a robust foundation for enhancing the utilization of cloud computing resources in the domain of task scheduling within a cloud computing environment.展开更多
Multispectral low earth orbit(LEO)satel-lites are characterized by a large volume of captured data and high spatial resolution,which can provide rich image information and data support for a vari-ety of fields,but it ...Multispectral low earth orbit(LEO)satel-lites are characterized by a large volume of captured data and high spatial resolution,which can provide rich image information and data support for a vari-ety of fields,but it is difficult for them to satisfy low-delay and low-energy consumed task processing re-quirements due to their limited computing resources.To address the above problems,this paper presents the LEO satellites cooperative task offloading and computing resource allocation(LEOC-TC)algorithm.Firstly,a LEO satellites cooperative task offloading system was designed so that the multispectral LEO satellites in the system could leave their tasks locally or offload them to other LEO satellites with servers for processing,thus providing high-quality information-processing services for multispectral LEO satellites.Secondly,an optimization problem with the objective of minimizing the weighted sum of the total task pro-cessing delay and total energy consumed for multi-spectral LEO satellite is established,and the optimiza-tion problem is split into an offloading ratio subprob-lem and a computing resource subproblem.Finally,Bernoulli mapping tuna swarm optimization algorithm is used to solve the above two sub-problems separately in order to satisfy the demand of low delay and low energy consumed by the system.Simulation results show that the total task processing cost of the LEOCTC algorithm can be reduced by 63.32%,66.67%,and 80.72%compared to the random offloading ratio algorithm,the average resource offloading algorithm,and the local computing algorithm,respectively.展开更多
基金supported in part by National Key R&D Program of China(Grant No.2022YFC3803700)in part by the National Natural Science Foundation of China(Grant No.92067102)in part by the project of Beijing Laboratory of Advanced Information Networks.
文摘The rise of time-sensitive applications with broad geographical scope drives the development of time-sensitive networking(TSN)from intra-domain to inter-domain to ensure overall end-to-end connectivity requirements in heterogeneous deployments.When multiple TSN networks interconnect over non-TSN networks,all devices in the network need to be syn-chronized by sharing a uniform time reference.How-ever,most non-TSN networks are best-effort.Path delay asymmetry and random noise accumulation can introduce unpredictable time errors during end-to-end time synchronization.These factors can degrade syn-chronization performance.Therefore,cross-domain time synchronization becomes a challenging issue for multiple TSN networks interconnected by non-TSN networks.This paper presents a cross-domain time synchronization scheme that follows the software-defined TSN(SD-TSN)paradigm.It utilizes a com-bined control plane constructed by a coordinate con-troller and a domain controller for centralized control and management of cross-domain time synchroniza-tion.The general operation flow of the cross-domain time synchronization process is designed.The mecha-nism of cross-domain time synchronization is revealed by introducing a synchronization model and an error compensation method.A TSN cross-domain proto-type testbed is constructed for verification.Results show that the scheme can achieve end-to-end high-precision time synchronization with accuracy and sta-bility.
基金this project under Geran Putra Inisiatif(GPI)with reference of GP-GPI/2023/976210。
文摘Accurate time synchronization is fundamental to the correct and efficient operation of Wireless Sensor Networks(WSNs),especially in security-critical,time-sensitive applications.However,most existing protocols degrade substantially under malicious interference.We introduce iSTSP,an Intelligent and Secure Time Synchronization Protocol that implements a four-stage defense pipeline to ensure robust,precise synchronization even in hostile environments:(1)trust preprocessing that filters node participation using behavioral trust scoring;(2)anomaly isolation employing a lightweight autoencoder to detect and excise malicious nodes in real time;(3)reliability-weighted consensus that prioritizes high-trust nodes during time aggregation;and(4)convergence-optimized synchronization that dynamically adjusts parameters using theoretical stability bounds.We provide rigorous convergence analysis including a closed-form expression for convergence time,and validate the protocol through both simulations and realworld experiments on a controlled 16-node testbed.Under Sybil attacks with five malicious nodes within this testbed,iSTSP maintains synchronization error increases under 12%and achieves a rapid convergence.Compared to state-ofthe-art protocols like TPSN,SE-FTSP,and MMAR-CTS,iSTSP offers 60%faster detection,broader threat coverage,and more than 7 times lower synchronization error,with a modest 9.3%energy overhead over 8 h.We argue this is an acceptable trade-off for mission-critical deployments requiring guaranteed security.These findings demonstrate iSTSP’s potential as a reliable solution for secure WSN synchronization and motivate future work on large-scale IoT deployments and integration with energy-efficient communication protocols.
基金supported by the Innovation Scientists and Technicians Troop Construction Projects of Henan Province(224000510002)。
文摘Time-Sensitive Network(TSN)with deterministic transmission capability is increasingly used in many emerging fields.It mainly guarantees the Quality of Service(QoS)of applications with strict requirements on time and security.One of the core features of TSN is traffic scheduling with bounded low delay in the network.However,traffic scheduling schemes in TSN are usually synthesized offline and lack dynamism.To implement incremental scheduling of newly arrived traffic in TSN,we propose a Dynamic Response Incremental Scheduling(DR-IS)method for time-sensitive traffic and deploy it on a software-defined time-sensitive network architecture.Under the premise of meeting the traffic scheduling requirements,we adopt two modes,traffic shift and traffic exchange,to dynamically adjust the time slot injection position of the traffic in the original scheme,and determine the sending offset time of the new timesensitive traffic to minimize the global traffic transmission jitter.The evaluation results show that DRIS method can effectively control the large increase of traffic transmission jitter in incremental scheduling without affecting the transmission delay,thus realizing the dynamic incremental scheduling of time-sensitive traffic in TSN.
基金funded by the Fundamental Research Funds for the Central Universities(J2023-024,J2023-027).
文摘As an important complement to cloud computing, edge computing can effectively reduce the workload of the backbone network. To reduce latency and energy consumption of edge computing, deep learning is used to learn the task offloading strategies by interacting with the entities. In actual application scenarios, users of edge computing are always changing dynamically. However, the existing task offloading strategies cannot be applied to such dynamic scenarios. To solve this problem, we propose a novel dynamic task offloading framework for distributed edge computing, leveraging the potential of meta-reinforcement learning (MRL). Our approach formulates a multi-objective optimization problem aimed at minimizing both delay and energy consumption. We model the task offloading strategy using a directed acyclic graph (DAG). Furthermore, we propose a distributed edge computing adaptive task offloading algorithm rooted in MRL. This algorithm integrates multiple Markov decision processes (MDP) with a sequence-to-sequence (seq2seq) network, enabling it to learn and adapt task offloading strategies responsively across diverse network environments. To achieve joint optimization of delay and energy consumption, we incorporate the non-dominated sorting genetic algorithm II (NSGA-II) into our framework. Simulation results demonstrate the superiority of our proposed solution, achieving a 21% reduction in time delay and a 19% decrease in energy consumption compared to alternative task offloading schemes. Moreover, our scheme exhibits remarkable adaptability, responding swiftly to changes in various network environments.
基金supported in part by the National Natural Science Foundation of China under Grant No.61473066in part by the Natural Science Foundation of Hebei Province under Grant No.F2021501020+2 种基金in part by the S&T Program of Qinhuangdao under Grant No.202401A195in part by the Science Research Project of Hebei Education Department under Grant No.QN2025008in part by the Innovation Capability Improvement Plan Project of Hebei Province under Grant No.22567637H
文摘Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power stations).To solve the problem,we propose an energy harvesting based task scheduling and resource management framework to provide robust and low-cost edge computing services for smart grid.First,we formulate an energy consumption minimization problem with regard to task offloading,time switching,and resource allocation for mobile devices,which can be decoupled and transformed into a typical knapsack problem.Then,solutions are derived by two different algorithms.Furthermore,we deploy renewable energy and energy storage units at edge servers to tackle intermittency and instability problems.Finally,we design an energy management algorithm based on sampling average approximation for edge computing servers to derive the optimal charging/discharging strategies,number of energy storage units,and renewable energy utilization.The simulation results show the efficiency and superiority of our proposed framework.
文摘With the rapid expansion of social media,analyzing emotions and their causes in texts has gained significant importance.Emotion-cause pair extraction enables the identification of causal relationships between emotions and their triggers within a text,facilitating a deeper understanding of expressed sentiments and their underlying reasons.This comprehension is crucial for making informed strategic decisions in various business and societal contexts.However,recent research approaches employing multi-task learning frameworks for modeling often face challenges such as the inability to simultaneouslymodel extracted features and their interactions,or inconsistencies in label prediction between emotion-cause pair extraction and independent assistant tasks like emotion and cause extraction.To address these issues,this study proposes an emotion-cause pair extraction methodology that incorporates joint feature encoding and task alignment mechanisms.The model consists of two primary components:First,joint feature encoding simultaneously generates features for emotion-cause pairs and clauses,enhancing feature interactions between emotion clauses,cause clauses,and emotion-cause pairs.Second,the task alignment technique is applied to reduce the labeling distance between emotion-cause pair extraction and the two assistant tasks,capturing deep semantic information interactions among tasks.The proposed method is evaluated on a Chinese benchmark corpus using 10-fold cross-validation,assessing key performance metrics such as precision,recall,and F1 score.Experimental results demonstrate that the model achieves an F1 score of 76.05%,surpassing the state-of-the-art by 1.03%.The proposed model exhibits significant improvements in emotion-cause pair extraction(ECPE)and cause extraction(CE)compared to existing methods,validating its effectiveness.This research introduces a novel approach based on joint feature encoding and task alignment mechanisms,contributing to advancements in emotion-cause pair extraction.However,the study’s limitation lies in the data sources,potentially restricting the generalizability of the findings.
基金supported by the Project of Science and Technology Research Program of Chongqing Education Commission of China(No.KJZD-K202401105)High-Quality Development Action Plan for Graduate Education at Chongqing University of Technology(No.gzljg2023308,No.gzljd2024204)+1 种基金the Graduate Innovation Program of Chongqing University of Technology(No.gzlcx20233197)Yunnan Provincial Key R&D Program(202203AA080006).
文摘Blockchain technology,based on decentralized data storage and distributed consensus design,has become a promising solution to address data security risks and provide privacy protection in the Internet-of-Things(IoT)due to its tamper-proof and non-repudiation features.Although blockchain typically does not require the endorsement of third-party trust organizations,it mostly needs to perform necessary mathematical calculations to prevent malicious attacks,which results in stricter requirements for computation resources on the participating devices.By offloading the computation tasks required to support blockchain consensus to edge service nodes or the cloud,while providing data privacy protection for IoT applications,it can effectively address the limitations of computation and energy resources in IoT devices.However,how to make reasonable offloading decisions for IoT devices remains an open issue.Due to the excellent self-learning ability of Reinforcement Learning(RL),this paper proposes a RL enabled Swarm Intelligence Optimization Algorithm(RLSIOA)that aims to improve the quality of initial solutions and achieve efficient optimization of computation task offloading decisions.The algorithm considers various factors that may affect the revenue obtained by IoT devices executing consensus algorithms(e.g.,Proof-of-Work),it optimizes the proportion of sub-tasks to be offloaded and the scale of computing resources to be rented from the edge and cloud to maximize the revenue of devices.Experimental results show that RLSIOA can obtain higher-quality offloading decision-making schemes at lower latency costs compared to representative benchmark algorithms.
基金supported and funded by theDeanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-RP23082).
文摘Fog computing is a key enabling technology of 6G systems as it provides quick and reliable computing,and data storage services which are required for several 6G applications.Artificial Intelligence(AI)algorithms will be an integral part of 6G systems and efficient task offloading techniques using fog computing will improve their performance and reliability.In this paper,the focus is on the scenario of Partial Offloading of a Task to Multiple Helpers(POMH)in which larger tasks are divided into smaller subtasks and processed in parallel,hence expediting task completion.However,using POMH presents challenges such as breaking tasks into subtasks and scaling these subtasks based on many interdependent factors to ensure that all subtasks of a task finish simultaneously,preventing resource wastage.Additionally,applying matching theory to POMH scenarios results in dynamic preference profiles of helping devices due to changing subtask sizes,resulting in a difficult-to-solve,externalities problem.This paper introduces a novel many-to-one matching-based algorithm,designed to address the externalities problem and optimize resource allocation within POMH scenarios.Additionally,we propose a new time-efficient preference profiling technique that further enhances time optimization in POMH scenarios.The performance of the proposed technique is thoroughly evaluated in comparison to alternate baseline schemes,revealing many advantages of the proposed approach.The simulation findings indisputably show that the proposed matching-based offloading technique outperforms existing methodologies in the literature,yielding a remarkable 52 reduction in task latency,particularly under high workloads.
基金supported by the Open Foundation of Henan Key Laboratory of Cyberspace Situation Awareness(No.HNTS2022020)the Science and Technology Research Program of Henan Province of China(232102210134,182102210130)Key Research Projects of Henan Provincial Universities(25B520005).
文摘With the development of vehicle networks and the construction of roadside units,Vehicular Ad Hoc Networks(VANETs)are increasingly promoting cooperative computing patterns among vehicles.Vehicular edge computing(VEC)offers an effective solution to mitigate resource constraints by enabling task offloading to edge cloud infrastructure,thereby reducing the computational burden on connected vehicles.However,this sharing-based and distributed computing paradigm necessitates ensuring the credibility and reliability of various computation nodes.Existing vehicular edge computing platforms have not adequately considered themisbehavior of vehicles.We propose a practical task offloading algorithm based on reputation assessment to address the task offloading problem in vehicular edge computing under an unreliable environment.This approach integrates deep reinforcement learning and reputation management to address task offloading challenges.Simulation experiments conducted using Veins demonstrate the feasibility and effectiveness of the proposed method.
文摘Drawing upon self-determination theory,this study examines the effects of vicarious abusive supervision on third-party’s self-efficacy and task performance within organizational contexts.Data were collected via surveys from 337 employees across diverse organizations.The results indicate that vicarious abusive supervision significantly undermines both self-efficacy and task performance among employees who are indirectly exposed to such behavior but not directly targeted.Furthermore,self-efficacy serves as a mediator between vicarious abusive supervision and task performance;however,this mediating effect is attenuated for employees with a high promotion focus.These findings provide valuable theoretical and practical insights,particularly in the domain of organizational behavior,by emphasizing the critical role of promotion focus in mitigating the negative effects of vicarious abusive supervision.This research contributes to the organizational behavior literature by shifting the focus from the traditional supervisor-subordinate dynamic to a third-party perspective,thereby enriching our understanding of how vicarious abusive supervision impacts employees within organizational settings.The study underscores the importance of self-efficacy and promotion focus as key factors in unethical leadership contexts.
基金supported by National Taiwan University Hospital Yunlin Branch Project NTUHYL 110.C018National Science and Technology Council,Taiwan.
文摘Porters play a crucial role in hospitals because they ensure the efficient transportation of patients,medical equipment,and vital documents.Despite its importance,there is a lack of research addressing the prediction of completion times for porter tasks.To address this gap,we utilized real-world porter delivery data from Taiwan University Hospital,China,Yunlin Branch,Taiwan Region of China.We first identified key features that can influence the duration of porter tasks.We then employed three widely-used machine learning algorithms:decision tree,random forest,and gradient boosting.To leverage the strengths of each algorithm,we finally adopted an ensemble modeling approach that aggregates their individual predictions.Our experimental results show that the proposed ensemble model can achieve a mean absolute error of 3 min in predicting task response time and 4.42 min in task completion time.The prediction error is around 50%lower compared to using only the historical average.These results demonstrate that our method significantly improves the accuracy of porter task time prediction,supporting better resource planning and patient care.It helps ward staff streamline workflows by reducing delays,enables porter managers to allocate resources more effectively,and shortens patient waiting times,contributing to a better care experience.
文摘Under the background of training practical compound talents in software engineering,this paper analyzes the problems existing in the current teaching of software engineering courses represented by software project management,puts forward the team task mechanism of software engineering courses with AI empowerment and cooperation and competition,develops a unified project management platform to support the implementation of team tasks,and proves the effectiveness of the scheme through the results obtained.
文摘The construction of ideological-political education within foreign language courses requires an integrated approach that encompasses value shaping,knowledge transfer,and competence cultivation.A critical challenge in this domain is the effective design and implementation of tasks that instill values,while also synergizing with acquiring knowledge and enhancing competencies.This paper delves into the philosophical underpinnings and operational principles of value-shaping task design and its practical application within the context of foreign language teaching.Utilizing Contemporary College English(Integrated Coursebook 3)as a case study,the paper explores value shaping-based task design in ideological-political education of foreign language courses,with the aim of providing references for the construction of ideological-political education in foreign language teaching.
基金The National Natural Science Foundation of China(62203015,62233001,62273351)The Beijing Natural Science Foundation(4242038)。
文摘This paper considers the swarm vigilance problem for multi-agent systems(MAS),where multiple agents are deployed within a rectangular region for perception-based vigilance.There are two main challenges,namely the task allocation for vigilance roles and the coverage planning of the perception ranges.Firstly,vigilance behavioral patterns and processes in animal populations within natural habitats are investigated.Inspired by these biological vigilance behaviors,an efficient vigilance task allocation model for MAS is proposed.Secondly,the subsequent optimization of task layouts can achieve efficient surveillance coverage with fewer agents,minimizing resource consumption.Thirdly,an improved particle swarm optimization(IPSO)algorithm is proposed,which incorporates fitness-driven adaptive inertia weight dynamics.According to simulation analysis and comparative studies,optimal parameter configurations for genetic algorithm(GA)and IPSO are determined.Finally,the results indicate the proposed IPSO's superior performance to both GA and standard particle swarm optimization(PSO)in vigilance task allocation optimization,with satisfying advantages in computational efficiency and solution quality.
基金funded by the National Key Research and Development Program of China under Grant 2019YFB1803301Beijing Natural Science Foundation (L202002)。
文摘Cybertwin-enabled 6th Generation(6G)network is envisioned to support artificial intelligence-native management to meet changing demands of 6G applications.Multi-Agent Deep Reinforcement Learning(MADRL)technologies driven by Cybertwins have been proposed for adaptive task offloading strategies.However,the existence of random transmission delay between Cybertwin-driven agents and underlying networks is not considered in related works,which destroys the standard Markov property and increases the decision reaction time to reduce the task offloading strategy performance.In order to address this problem,we propose a pipelining task offloading method to lower the decision reaction time and model it as a delay-aware Markov Decision Process(MDP).Then,we design a delay-aware MADRL algorithm to minimize the weighted sum of task execution latency and energy consumption.Firstly,the state space is augmented using the lastly-received state and historical actions to rebuild the Markov property.Secondly,Gate Transformer-XL is introduced to capture historical actions'importance and maintain the consistent input dimension dynamically changed due to random transmission delays.Thirdly,a sampling method and a new loss function with the difference between the current and target state value and the difference between real state-action value and augmented state-action value are designed to obtain state transition trajectories close to the real ones.Numerical results demonstrate that the proposed methods are effective in reducing reaction time and improving the task offloading performance in the random-delay Cybertwin-enabled 6G networks.
基金the Deanship of Postgraduate Studies and Scientific Research at Majmaah University for funding this research work through the project number(R-2025-1567).
文摘Due to the intense data flow in expanding Internet of Things(IoT)applications,a heavy processing cost and workload on the fog-cloud side become inevitable.One of the most critical challenges is optimal task scheduling.Since this is an NP-hard problem type,a metaheuristic approach can be a good option.This study introduces a novel enhancement to the Artificial Rabbits Optimization(ARO)algorithm by integrating Chaotic maps and Levy flight strategies(CLARO).This dual approach addresses the limitations of standard ARO in terms of population diversity and convergence speed.It is designed for task scheduling in fog-cloud environments,optimizing energy consumption,makespan,and execution time simultaneously three critical parameters often treated individually in prior works.Unlike conventional single-objective methods,the proposed approach incorporates a multi-objective fitness function that dynamically adjusts the weight of each parameter,resulting in better resource allocation and load balancing.In analysis,a real-world dataset,the Open-source Google Cloud Jobs Dataset(GoCJ_Dataset),is used for performance measurement,and analyses are performed on three considered parameters.Comparisons are applied with well-known algorithms:GWO,SCSO,PSO,WOA,and ARO to indicate the reliability of the proposed method.In this regard,performance evaluation is performed by assigning these tasks to Virtual Machines(VMs)in the resource pool.Simulations are performed on 90 base cases and 30 scenarios for each evaluation parameter.The results indicated that the proposed algorithm achieved the best makespan performance in 80% of cases,ranked first in execution time in 61%of cases,and performed best in the final parameter in 69% of cases.In addition,according to the obtained results based on the defined fitness function,the proposed method(CLARO)is 2.52%better than ARO,3.95%better than SCSO,5.06%better than GWO,8.15%better than PSO,and 9.41%better than WOA.
基金Supported by Wuxi Municipal Health Commission Major Project,No.Z202107Wuxi Taihu Talent Project,No.WXTTP 2021.
文摘BACKGROUND The neural mechanisms underlying aggressive behavior in schizophrenia(SCZ)remain poorly understood.To date,no studies have reported on the event-related potential(ERP)characteristics of aggression in SCZ using the competitive reaction time task(CRTT).Further investigation into the ERP correlates of aggression in SCZ would provide valuable insights into the neural processes involved.AIM To explore the neural mechanism of aggressive behavior in SCZ.METHODS Participants of this study included 40 SCZ patients and 42 healthy controls(HCs).The Reactive Proactive Aggression Questionnaire was used to assess trait of aggression.The Barratt Impulsiveness Scale 11 was used to measure impulsiveness.The Positive and Negative Symptom Scale(PANSS)was used to evaluate psychopathological features and disease severity.All participants were measured with ERP while performing the CRTT.Data of behavior,ERP components(P2,N2,and P3),and feedback-related negativity(FRN)were analyzed.RESULTS Analysis of the behavioral data revealed that compared with HCs,SCZ patients exhibited higher punishment choices.Analysis of ERP components showed that compared with HCs,SCZ patients exhibited higher N2 amplitudes and P2 amplitudes during the decision phase of the CRTT;however,SCZ patients exhibited lower FRN amplitudes and lower P3 amplitudes during the outcome phase of the CRTT.The N2 amplitudes evoked by highintensity provocation were positively related to PANSS-P scores.And the P3 amplitudes evoked in the winning trials were negatively correlated with the PANSS-G scores.CONCLUSION SCZ patients exhibit abnormal ERP characteristics evoked by the CRTT,which suggests the neural correlates of aggressive behavior in SCZ.
基金supported by the National Natural Science Foundation of China(Grant No.62103349)the Henan Province Science and Technology Research Project(Grant No.232102210104).
文摘Edge computing has transformed smart grids by lowering latency,reducing network congestion,and enabling real-time decision-making.Nevertheless,devising an optimal task-offloading strategy remains challenging,as it must jointly minimise energy consumption and response time under fluctuating workloads and volatile network conditions.We cast the offloading problem as aMarkov Decision Process(MDP)and solve it with Deep Reinforcement Learning(DRL).Specifically,we present a three-tier architecture—end devices,edge nodes,and a cloud server—and enhance Proximal Policy Optimization(PPO)to learn adaptive,energy-aware policies.A Convolutional Neural Network(CNN)extracts high-level features from system states,enabling the agent to respond continually to changing conditions.Extensive simulations show that the proposed method reduces task latency and energy consumption far more than several baseline algorithms,thereby improving overall system performance.These results demonstrate the effectiveness and robustness of the framework for real-time task offloading in dynamic smart-grid environments.
文摘The widespread adoption of cloud computing has underscored the critical importance of efficient resource allocation and management, particularly in task scheduling, which involves assigning tasks to computing resources for optimized resource utilization. Several meta-heuristic algorithms have shown effectiveness in task scheduling, among which the relatively recent Willow Catkin Optimization (WCO) algorithm has demonstrated potential, albeit with apparent needs for enhanced global search capability and convergence speed. To address these limitations of WCO in cloud computing task scheduling, this paper introduces an improved version termed the Advanced Willow Catkin Optimization (AWCO) algorithm. AWCO enhances the algorithm’s performance by augmenting its global search capability through a quasi-opposition-based learning strategy and accelerating its convergence speed via sinusoidal mapping. A comprehensive evaluation utilizing the CEC2014 benchmark suite, comprising 30 test functions, demonstrates that AWCO achieves superior optimization outcomes, surpassing conventional WCO and a range of established meta-heuristics. The proposed algorithm also considers trade-offs among the cost, makespan, and load balancing objectives. Experimental results of AWCO are compared with those obtained using the other meta-heuristics, illustrating that the proposed algorithm provides superior performance in task scheduling. The method offers a robust foundation for enhancing the utilization of cloud computing resources in the domain of task scheduling within a cloud computing environment.
基金supported in part by Sub Project of National Key Research and Development plan in 2020(No.2020YFC1511704)scientific research level improvement project to promote the colleges connotation development of Beijing Information Science&Technology University(No.2020KYNH212,No.2021CGZH302)in part by the National Natural Science Foundation of China(Grant No.61971048).
文摘Multispectral low earth orbit(LEO)satel-lites are characterized by a large volume of captured data and high spatial resolution,which can provide rich image information and data support for a vari-ety of fields,but it is difficult for them to satisfy low-delay and low-energy consumed task processing re-quirements due to their limited computing resources.To address the above problems,this paper presents the LEO satellites cooperative task offloading and computing resource allocation(LEOC-TC)algorithm.Firstly,a LEO satellites cooperative task offloading system was designed so that the multispectral LEO satellites in the system could leave their tasks locally or offload them to other LEO satellites with servers for processing,thus providing high-quality information-processing services for multispectral LEO satellites.Secondly,an optimization problem with the objective of minimizing the weighted sum of the total task pro-cessing delay and total energy consumed for multi-spectral LEO satellite is established,and the optimiza-tion problem is split into an offloading ratio subprob-lem and a computing resource subproblem.Finally,Bernoulli mapping tuna swarm optimization algorithm is used to solve the above two sub-problems separately in order to satisfy the demand of low delay and low energy consumed by the system.Simulation results show that the total task processing cost of the LEOCTC algorithm can be reduced by 63.32%,66.67%,and 80.72%compared to the random offloading ratio algorithm,the average resource offloading algorithm,and the local computing algorithm,respectively.