With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on comput...With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on computing latency and resource supply efficiency.Multi-access edge computing technology deploys cloud computing capabilities at the network edge;constructs distributed computing nodes and multi-access systems and offers infrastructure support for services with low latency and high reliability.Existing research relies on a strong assumption that the environmental state is fully observable and fails to thoroughly consider the continuous time-varying features of edge server load fluctuations,leading to insufficient adaptability of the model in a heterogeneous dynamic environment.Thus,this paper establishes a framework for end-edge collaborative task offloading based on a partially observable Markov decision-making process(POMDP)and proposes a method for end-edge collaborative task offloading in heterogeneous scenarios.It achieves time-series modeling of the historical load characteristics of edge servers and endows the agent with the ability to be aware of the load in dynamic environmental states.Moreover,by dynamically assessing the exploration value of historical trajectories in the central trajectory pool and adjusting the sample weight distribution,directional exploration and strategy optimization of high-value trajectories are realized.Experimental results indicate that the proposed method exhibits distinct advantages compared with existing methods in terms of average delay and task failure rate and also verifies the method’s robustness in a dynamic environment.展开更多
The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing r...The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing resources has led to a scarcity.In order to meet the needs of an increasing number of researchers,it is imperative to facilitate efficient and flexible access to computing resources in a cloud environment.In this paper,we propose a novel quantum computing paradigm,Virtual QPU(VQPU),which addresses this issue and enhances quantum cloud throughput with guaranteed circuit fidelity.The proposal introduces three innovative concepts:(1)The integration of virtualization technology into the field of quantum computing to enhance quantum cloud throughput.(2)The introduction of an asynchronous execution of circuits methodology to improve quantum computing flexibility.(3)The development of a virtual QPU allocation scheme for quantum tasks in a cloud environment to improve circuit fidelity.The concepts have been validated through the utilization of a self-built simulated quantum cloud platform.展开更多
In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem i...In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem in grid computing. It first generates a new group of individuals through genetic operation such as reproduction, crossover, mutation, etc, and than simulated anneals independently all the generated individuals respectively. When the temperature in the process of cooling no longer falls, the result is the optimal solution on the whole. From the analysis and experiment result, it is concluded that this algorithm is superior to genetic algorithm and simulated annealing.展开更多
Offloading application to cloud can augment mobile devices' computation capabilities for the emerging resource-hungry mobile application, however it can also consume both much time and energy for mobile device off...Offloading application to cloud can augment mobile devices' computation capabilities for the emerging resource-hungry mobile application, however it can also consume both much time and energy for mobile device offloading application remotely to cloud. In this paper, we develop a newly adaptive application offloading decision-transmission scheduling scheme which can solve above problem efficiently. Specifically, we first propose an adaptive application offloading model which allows multiple target clouds coexisting. Second, based on Lyapunov optimization theory, a low complexity adaptive offloading decision-transmission scheduling scheme has been proposed. And the performance analysis is also given. Finally, simulation results show that,compared with that all applications are executed locally, mobile device can save 68.557% average execution time and 67.095% average energy consumption under situations.展开更多
Many Task Computing(MTC)is a new class of computing paradigm in which the aggregate number of tasks,quantity of computing,and volumes of data may be extremely large.With the advent of Cloud computing and big data era,...Many Task Computing(MTC)is a new class of computing paradigm in which the aggregate number of tasks,quantity of computing,and volumes of data may be extremely large.With the advent of Cloud computing and big data era,scheduling and executing large-scale computing tasks efficiently and allocating resources to tasks reasonably are becoming a quite challenging problem.To improve both task execution and resource utilization efficiency,we present a task scheduling algorithm with resource attribute selection,which can select the optimal node to execute a task according to its resource requirements and the fitness between the resource node and the task.Experiment results show that there is significant improvement in execution throughput and resource utilization compared with the other three algorithms and four scheduling frameworks.In the scheduling algorithm comparison,the throughput is 77%higher than Min-Min algorithm and the resource utilization can reach 91%.In the scheduling framework comparison,the throughput(with work-stealing)is at least 30%higher than the other frameworks and the resource utilization reaches 94%.The scheduling algorithm can make a good model for practical MTC applications.展开更多
With the rapid development of cloud computing,edge computing,and smart devices,computing power resources indicate a trend of ubiquitous deployment.The traditional network architecture cannot efficiently leverage these...With the rapid development of cloud computing,edge computing,and smart devices,computing power resources indicate a trend of ubiquitous deployment.The traditional network architecture cannot efficiently leverage these distributed computing power resources due to computing power island effect.To overcome these problems and improve network efficiency,a new network computing paradigm is proposed,i.e.,Computing Power Network(CPN).Computing power network can connect ubiquitous and heterogenous computing power resources through networking to realize computing power scheduling flexibly.In this survey,we make an exhaustive review on the state-of-the-art research efforts on computing power network.We first give an overview of computing power network,including definition,architecture,and advantages.Next,a comprehensive elaboration of issues on computing power modeling,information awareness and announcement,resource allocation,network forwarding,computing power transaction platform and resource orchestration platform is presented.The computing power network testbed is built and evaluated.The applications and use cases in computing power network are discussed.Then,the key enabling technologies for computing power network are introduced.Finally,open challenges and future research directions are presented as well.展开更多
In traditional wireless broadcast networks,a corrupted packet must be retransmitted even if it has been lost by only one receiver.Obviously,this is not bandwidth-efficient for the receivers that already hold the retra...In traditional wireless broadcast networks,a corrupted packet must be retransmitted even if it has been lost by only one receiver.Obviously,this is not bandwidth-efficient for the receivers that already hold the retransmitted packet.Therefore,it is important to develop a method to realise efficient broadcast transmission.Network coding is a promising technique in this scenario.However,none of the proposed schemes achieves both high transmission efficiency and low computational complexity simultaneously so far.To address this problem,a novel Efficient Opportunistic Network Coding Retransmission(EONCR)scheme is proposed in this paper.This scheme employs a new packet scheduling algorithm which uses a Packet Distribution Matrix(PDM)directly to select the coded packets.The analysis and simulation results indicate that transmission efficiency of EONCR is over 0.1,more than the schemes proposed previously in some simulation conditions,and the computational overhead is reduced substantially.Hence,it has great application prospects in wireless broadcast networks,especially energyand bandwidth-limited systems such as satellite broadcast systems and Planetary Networks(PNs).展开更多
Load-time series data in mobile cloud computing of Internet of Vehicles(IoV)usually have linear and nonlinear composite characteristics.In order to accurately describe the dynamic change trend of such loads,this study...Load-time series data in mobile cloud computing of Internet of Vehicles(IoV)usually have linear and nonlinear composite characteristics.In order to accurately describe the dynamic change trend of such loads,this study designs a load prediction method by using the resource scheduling model for mobile cloud computing of IoV.Firstly,a chaotic analysis algorithm is implemented to process the load-time series,while some learning samples of load prediction are constructed.Secondly,a support vector machine(SVM)is used to establish a load prediction model,and an improved artificial bee colony(IABC)function is designed to enhance the learning ability of the SVM.Finally,a CloudSim simulation platform is created to select the perminute CPU load history data in the mobile cloud computing system,which is composed of 50 vehicles as the data set;and a comparison experiment is conducted by using a grey model,a back propagation neural network,a radial basis function(RBF)neural network and a RBF kernel function of SVM.As shown in the experimental results,the prediction accuracy of the method proposed in this study is significantly higher than other models,with a significantly reduced real-time prediction error for resource loading in mobile cloud environments.Compared with single-prediction models,the prediction method proposed can build up multidimensional time series in capturing complex load time series,fit and describe the load change trends,approximate the load time variability more precisely,and deliver strong generalization ability to load prediction models for mobile cloud computing resources.展开更多
Today, most construction projects in urban environments are complex high-rise buildings that present unique challenges, including local building ordinances and restrictions, adjoining public and residential areas, nar...Today, most construction projects in urban environments are complex high-rise buildings that present unique challenges, including local building ordinances and restrictions, adjoining public and residential areas, narrow sidewalks and streets, and underground utilities, all of which require extensive planning and tight schedules. A major problem facing such projects is to formulate realistic schedules that will make it possible to meet contractual completion dates with limited resources and budgets. The scheduling software products currently used in construction projects, which include Primavera P6, Microsoft Project, etc., are not actually applied as a scheduling tool in practical construction projects, which instead generally depend on Microsoft Excel or a bar-chart. This is because the existing scheduling programs cannot provide more user-oriented schedule format such as representing two-way multiple overlapping relationships. To overcome this deficiency, the BDM (beeline diagramming method) is proposed as a new networking technique in 2010. But two-way multiple overlapping relationships generate the loop in a conventional schedule computation process. This paper addresses the loop phenomenon of two-way multiple overlapping relationships in a BDM network as well as proposes the solutions of them, and then presents a practical application of two-way multiple overlapping relationships at a real project.展开更多
High energy consumption is one of the key issues of cloud computing systems. Incoming jobs in cloud computing environments have the nature of randomness, and compute nodes have to be powered on all the time to await i...High energy consumption is one of the key issues of cloud computing systems. Incoming jobs in cloud computing environments have the nature of randomness, and compute nodes have to be powered on all the time to await incoming tasks. This results in a great waste of energy. An energy-saving task scheduling algorithm based on the vacation queuing model for cloud computing systems is proposed in this paper. First, we use the vacation queuing model with exhaustive service to model the task schedule of a heterogeneous cloud computing system.Next, based on the busy period and busy cycle under steady state, we analyze the expectations of task sojourn time and energy consumption of compute nodes in the heterogeneous cloud computing system. Subsequently, we propose a task scheduling algorithm based on similar tasks to reduce the energy consumption. Simulation results show that the proposed algorithm can reduce the energy consumption of the cloud computing system effectively while meeting the task performance.展开更多
基金funded by the State Grid Corporation Science and Technology Project“Research and Application of Key Technologies for Integrated Sensing and Computing for Intelligent Operation of Power Grid”(Grant No.5700-202318596A-3-2-ZN).
文摘With the rapid development of power Internet of Things(IoT)scenarios such as smart factories and smart homes,numerous intelligent terminal devices and real-time interactive applications impose higher demands on computing latency and resource supply efficiency.Multi-access edge computing technology deploys cloud computing capabilities at the network edge;constructs distributed computing nodes and multi-access systems and offers infrastructure support for services with low latency and high reliability.Existing research relies on a strong assumption that the environmental state is fully observable and fails to thoroughly consider the continuous time-varying features of edge server load fluctuations,leading to insufficient adaptability of the model in a heterogeneous dynamic environment.Thus,this paper establishes a framework for end-edge collaborative task offloading based on a partially observable Markov decision-making process(POMDP)and proposes a method for end-edge collaborative task offloading in heterogeneous scenarios.It achieves time-series modeling of the historical load characteristics of edge servers and endows the agent with the ability to be aware of the load in dynamic environmental states.Moreover,by dynamically assessing the exploration value of historical trajectories in the central trajectory pool and adjusting the sample weight distribution,directional exploration and strategy optimization of high-value trajectories are realized.Experimental results indicate that the proposed method exhibits distinct advantages compared with existing methods in terms of average delay and task failure rate and also verifies the method’s robustness in a dynamic environment.
文摘The increasing popularity of quantum computing has resulted in a considerable rise in demand for cloud quantum computing usage in recent years.Nevertheless,the rapid surge in demand for cloud-based quantum computing resources has led to a scarcity.In order to meet the needs of an increasing number of researchers,it is imperative to facilitate efficient and flexible access to computing resources in a cloud environment.In this paper,we propose a novel quantum computing paradigm,Virtual QPU(VQPU),which addresses this issue and enhances quantum cloud throughput with guaranteed circuit fidelity.The proposal introduces three innovative concepts:(1)The integration of virtualization technology into the field of quantum computing to enhance quantum cloud throughput.(2)The introduction of an asynchronous execution of circuits methodology to improve quantum computing flexibility.(3)The development of a virtual QPU allocation scheme for quantum tasks in a cloud environment to improve circuit fidelity.The concepts have been validated through the utilization of a self-built simulated quantum cloud platform.
基金Supported by the National Basic ResearchProgramof China (973 Program2003CB314804)
文摘In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem in grid computing. It first generates a new group of individuals through genetic operation such as reproduction, crossover, mutation, etc, and than simulated anneals independently all the generated individuals respectively. When the temperature in the process of cooling no longer falls, the result is the optimal solution on the whole. From the analysis and experiment result, it is concluded that this algorithm is superior to genetic algorithm and simulated annealing.
基金supported by National Natural Science Foundation of China (Grant No.61261017, No.61571143 and No.61561014)Guangxi Natural Science Foundation (2013GXNSFAA019334 and 2014GXNSFAA118387)+3 种基金Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education (No.CRKL150112)Guangxi Key Lab of Wireless Wideband Communication & Signal Processing (GXKL0614202, GXKL0614101 and GXKL061501)Sci.and Tech.on Info.Transmission and Dissemination in Communication Networks Lab (No.ITD-U14008/KX142600015)Graduate Student Research Innovation Project of Guilin University of Electronic Technology (YJCXS201523)
文摘Offloading application to cloud can augment mobile devices' computation capabilities for the emerging resource-hungry mobile application, however it can also consume both much time and energy for mobile device offloading application remotely to cloud. In this paper, we develop a newly adaptive application offloading decision-transmission scheduling scheme which can solve above problem efficiently. Specifically, we first propose an adaptive application offloading model which allows multiple target clouds coexisting. Second, based on Lyapunov optimization theory, a low complexity adaptive offloading decision-transmission scheduling scheme has been proposed. And the performance analysis is also given. Finally, simulation results show that,compared with that all applications are executed locally, mobile device can save 68.557% average execution time and 67.095% average energy consumption under situations.
基金ACKNOWLEDGEMENTS The authors would like to thank the reviewers for their detailed reviews and constructive comments, which have helped improve the quality of this paper. The research has been partly supported by National Natural Science Foundation of China No. 61272528 and No. 61034005, and the Central University Fund (ID-ZYGX2013J073).
文摘Many Task Computing(MTC)is a new class of computing paradigm in which the aggregate number of tasks,quantity of computing,and volumes of data may be extremely large.With the advent of Cloud computing and big data era,scheduling and executing large-scale computing tasks efficiently and allocating resources to tasks reasonably are becoming a quite challenging problem.To improve both task execution and resource utilization efficiency,we present a task scheduling algorithm with resource attribute selection,which can select the optimal node to execute a task according to its resource requirements and the fitness between the resource node and the task.Experiment results show that there is significant improvement in execution throughput and resource utilization compared with the other three algorithms and four scheduling frameworks.In the scheduling algorithm comparison,the throughput is 77%higher than Min-Min algorithm and the resource utilization can reach 91%.In the scheduling framework comparison,the throughput(with work-stealing)is at least 30%higher than the other frameworks and the resource utilization reaches 94%.The scheduling algorithm can make a good model for practical MTC applications.
基金supported by the National Science Foundation of China under Grant 62271062 and 62071063by the Zhijiang Laboratory Open Project Fund 2020LCOAB01。
文摘With the rapid development of cloud computing,edge computing,and smart devices,computing power resources indicate a trend of ubiquitous deployment.The traditional network architecture cannot efficiently leverage these distributed computing power resources due to computing power island effect.To overcome these problems and improve network efficiency,a new network computing paradigm is proposed,i.e.,Computing Power Network(CPN).Computing power network can connect ubiquitous and heterogenous computing power resources through networking to realize computing power scheduling flexibly.In this survey,we make an exhaustive review on the state-of-the-art research efforts on computing power network.We first give an overview of computing power network,including definition,architecture,and advantages.Next,a comprehensive elaboration of issues on computing power modeling,information awareness and announcement,resource allocation,network forwarding,computing power transaction platform and resource orchestration platform is presented.The computing power network testbed is built and evaluated.The applications and use cases in computing power network are discussed.Then,the key enabling technologies for computing power network are introduced.Finally,open challenges and future research directions are presented as well.
基金supported in part by the National Natural Science Foundation of China under Grant No. 61032004the National High Technical Research and Development Program of China (863 Program) under Grants No. 2012AA121605,No. 2012AA01A503,No.2012AA01A510
文摘In traditional wireless broadcast networks,a corrupted packet must be retransmitted even if it has been lost by only one receiver.Obviously,this is not bandwidth-efficient for the receivers that already hold the retransmitted packet.Therefore,it is important to develop a method to realise efficient broadcast transmission.Network coding is a promising technique in this scenario.However,none of the proposed schemes achieves both high transmission efficiency and low computational complexity simultaneously so far.To address this problem,a novel Efficient Opportunistic Network Coding Retransmission(EONCR)scheme is proposed in this paper.This scheme employs a new packet scheduling algorithm which uses a Packet Distribution Matrix(PDM)directly to select the coded packets.The analysis and simulation results indicate that transmission efficiency of EONCR is over 0.1,more than the schemes proposed previously in some simulation conditions,and the computational overhead is reduced substantially.Hence,it has great application prospects in wireless broadcast networks,especially energyand bandwidth-limited systems such as satellite broadcast systems and Planetary Networks(PNs).
基金This work was supported by Shandong medical and health science and technology development plan project(No.202012070393).
文摘Load-time series data in mobile cloud computing of Internet of Vehicles(IoV)usually have linear and nonlinear composite characteristics.In order to accurately describe the dynamic change trend of such loads,this study designs a load prediction method by using the resource scheduling model for mobile cloud computing of IoV.Firstly,a chaotic analysis algorithm is implemented to process the load-time series,while some learning samples of load prediction are constructed.Secondly,a support vector machine(SVM)is used to establish a load prediction model,and an improved artificial bee colony(IABC)function is designed to enhance the learning ability of the SVM.Finally,a CloudSim simulation platform is created to select the perminute CPU load history data in the mobile cloud computing system,which is composed of 50 vehicles as the data set;and a comparison experiment is conducted by using a grey model,a back propagation neural network,a radial basis function(RBF)neural network and a RBF kernel function of SVM.As shown in the experimental results,the prediction accuracy of the method proposed in this study is significantly higher than other models,with a significantly reduced real-time prediction error for resource loading in mobile cloud environments.Compared with single-prediction models,the prediction method proposed can build up multidimensional time series in capturing complex load time series,fit and describe the load change trends,approximate the load time variability more precisely,and deliver strong generalization ability to load prediction models for mobile cloud computing resources.
文摘Today, most construction projects in urban environments are complex high-rise buildings that present unique challenges, including local building ordinances and restrictions, adjoining public and residential areas, narrow sidewalks and streets, and underground utilities, all of which require extensive planning and tight schedules. A major problem facing such projects is to formulate realistic schedules that will make it possible to meet contractual completion dates with limited resources and budgets. The scheduling software products currently used in construction projects, which include Primavera P6, Microsoft Project, etc., are not actually applied as a scheduling tool in practical construction projects, which instead generally depend on Microsoft Excel or a bar-chart. This is because the existing scheduling programs cannot provide more user-oriented schedule format such as representing two-way multiple overlapping relationships. To overcome this deficiency, the BDM (beeline diagramming method) is proposed as a new networking technique in 2010. But two-way multiple overlapping relationships generate the loop in a conventional schedule computation process. This paper addresses the loop phenomenon of two-way multiple overlapping relationships in a BDM network as well as proposes the solutions of them, and then presents a practical application of two-way multiple overlapping relationships at a real project.
基金supported by Research and Innovation Projects for Graduates of Jiangsu Graduates of Jiangsu Province (No. CXZZ12 0483)the Science and Technology Support Program of Jiangsu Province (No. BE2012849)
文摘High energy consumption is one of the key issues of cloud computing systems. Incoming jobs in cloud computing environments have the nature of randomness, and compute nodes have to be powered on all the time to await incoming tasks. This results in a great waste of energy. An energy-saving task scheduling algorithm based on the vacation queuing model for cloud computing systems is proposed in this paper. First, we use the vacation queuing model with exhaustive service to model the task schedule of a heterogeneous cloud computing system.Next, based on the busy period and busy cycle under steady state, we analyze the expectations of task sojourn time and energy consumption of compute nodes in the heterogeneous cloud computing system. Subsequently, we propose a task scheduling algorithm based on similar tasks to reduce the energy consumption. Simulation results show that the proposed algorithm can reduce the energy consumption of the cloud computing system effectively while meeting the task performance.