期刊文献+
共找到113篇文章
< 1 2 6 >
每页显示 20 50 100
Parallel Test Tasks Scheduling and Resources Configuration Based on GA-ACA 被引量:3
1
作者 方甲永 薛辉辉 肖明清 《Journal of Measurement Science and Instrumentation》 CAS 2011年第4期321-326,共6页
A Genetic Algorithm-Ant Colony Algorithm(GA-ACA),which can be used to optimize multi-Unit Under Test(UUT)parallel test tasks sequences and resources configuration quickly and accurately,is proposed in the paper.With t... A Genetic Algorithm-Ant Colony Algorithm(GA-ACA),which can be used to optimize multi-Unit Under Test(UUT)parallel test tasks sequences and resources configuration quickly and accurately,is proposed in the paper.With the establishment of the mathematic model of multi-UUT parallel test tasks and resources,the condition of multi-UUT resources mergence is analyzed to obtain minimum resource requirement under minimum test time.The definition of cost efficiency is put forward,followed by the design of gene coding and path selection project,which can satisfy multi-UUT parallel test tasks scheduling.At the threshold of the algorithm,GA is adopted to provide initial pheromone for ACA,and then dual-convergence pheromone feedback mode is applied in ACA to avoid local optimization and parameters dependence.The practical application proves that the algorithm has a remarkable effect on solving the problems of multi-UUT parallel test tasks scheduling and resources configuration. 展开更多
关键词 parallel test Genetic Algorithm-Ant Colony Algo-rithm GA-ACA cost efficiency multi-UnitUnder Test UUT resources configuration tasks scheduling
在线阅读 下载PDF
Tasks Scheduling in Cloud Environment Using PSO-BATS with MLRHE
2
作者 Anwar R Shaheen Sundar Santhosh Kumar 《Intelligent Automation & Soft Computing》 SCIE 2023年第3期2963-2978,共16页
Cloud computing plays a significant role in Information Technology(IT)industry to deliver scalable resources as a service.One of the most important factor to increase the performance of the cloud server is maximizing t... Cloud computing plays a significant role in Information Technology(IT)industry to deliver scalable resources as a service.One of the most important factor to increase the performance of the cloud server is maximizing the resource utilization in task scheduling.The main advantage of this scheduling is to max-imize the performance and minimize the time loss.Various researchers examined numerous scheduling methods to achieve Quality of Service(QoS)and to reduce execution time.However,it had disadvantages in terms of low throughput and high response time.Hence,this study aimed to schedule the task efficiently and to eliminate the faults in scheduling the tasks to the Virtual Machines(VMs).For this purpose,the research proposed novel Particle Swarm Optimization-Bandwidth Aware divisible Task(PSO-BATS)scheduling with Multi-Layered Regression Host Employment(MLRHE)to sort out the issues of task scheduling and ease the scheduling operation by load balancing.The proposed efficient sche-duling provides benefits to both cloud users and servers.The performance evalua-tion is undertaken with respect to cost,Performance Improvement Rate(PIR)and makespan which revealed the efficiency of the proposed method.Additionally,comparative analysis is undertaken which confirmed the performance of the intro-duced system than conventional system for scheduling tasks with highflexibility. 展开更多
关键词 Task scheduling virtual machines(VM) particle swarm optimization(PSO) bandwidth aware divisible task scheduling(BATS) multi-layered regression
在线阅读 下载PDF
Providing Robust and Low-Cost Edge Computing in Smart Grid:An Energy Harvesting Based Task Scheduling and Resource Management Framework 被引量:1
3
作者 Xie Zhigang Song Xin +1 位作者 Xu Siyang Cao Jing 《China Communications》 2025年第2期226-240,共15页
Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power sta... Recently,one of the main challenges facing the smart grid is insufficient computing resources and intermittent energy supply for various distributed components(such as monitoring systems for renewable energy power stations).To solve the problem,we propose an energy harvesting based task scheduling and resource management framework to provide robust and low-cost edge computing services for smart grid.First,we formulate an energy consumption minimization problem with regard to task offloading,time switching,and resource allocation for mobile devices,which can be decoupled and transformed into a typical knapsack problem.Then,solutions are derived by two different algorithms.Furthermore,we deploy renewable energy and energy storage units at edge servers to tackle intermittency and instability problems.Finally,we design an energy management algorithm based on sampling average approximation for edge computing servers to derive the optimal charging/discharging strategies,number of energy storage units,and renewable energy utilization.The simulation results show the efficiency and superiority of our proposed framework. 展开更多
关键词 edge computing energy harvesting energy storage unit renewable energy sampling average approximation task scheduling
在线阅读 下载PDF
A Bioinspired Method for Optimal Task Scheduling in Fog-Cloud Environment
4
作者 Ferzat Anka Ghanshyam G.Tejani +1 位作者 Sunil Kumar Sharma Mohammed Baljon 《Computer Modeling in Engineering & Sciences》 2025年第3期2691-2724,共34页
Due to the intense data flow in expanding Internet of Things(IoT)applications,a heavy processing cost and workload on the fog-cloud side become inevitable.One of the most critical challenges is optimal task scheduling... Due to the intense data flow in expanding Internet of Things(IoT)applications,a heavy processing cost and workload on the fog-cloud side become inevitable.One of the most critical challenges is optimal task scheduling.Since this is an NP-hard problem type,a metaheuristic approach can be a good option.This study introduces a novel enhancement to the Artificial Rabbits Optimization(ARO)algorithm by integrating Chaotic maps and Levy flight strategies(CLARO).This dual approach addresses the limitations of standard ARO in terms of population diversity and convergence speed.It is designed for task scheduling in fog-cloud environments,optimizing energy consumption,makespan,and execution time simultaneously three critical parameters often treated individually in prior works.Unlike conventional single-objective methods,the proposed approach incorporates a multi-objective fitness function that dynamically adjusts the weight of each parameter,resulting in better resource allocation and load balancing.In analysis,a real-world dataset,the Open-source Google Cloud Jobs Dataset(GoCJ_Dataset),is used for performance measurement,and analyses are performed on three considered parameters.Comparisons are applied with well-known algorithms:GWO,SCSO,PSO,WOA,and ARO to indicate the reliability of the proposed method.In this regard,performance evaluation is performed by assigning these tasks to Virtual Machines(VMs)in the resource pool.Simulations are performed on 90 base cases and 30 scenarios for each evaluation parameter.The results indicated that the proposed algorithm achieved the best makespan performance in 80% of cases,ranked first in execution time in 61%of cases,and performed best in the final parameter in 69% of cases.In addition,according to the obtained results based on the defined fitness function,the proposed method(CLARO)is 2.52%better than ARO,3.95%better than SCSO,5.06%better than GWO,8.15%better than PSO,and 9.41%better than WOA. 展开更多
关键词 Improved ARO fog computing task scheduling GoCJ_Dataset chaotic map levy flight
在线阅读 下载PDF
Innovative Approaches to Task Scheduling in Cloud Computing Environments Using an Advanced Willow Catkin Optimization Algorithm
5
作者 Jeng-Shyang Pan Na Yu +3 位作者 Shu-Chuan Chu An-Ning Zhang Bin Yan Junzo Watada 《Computers, Materials & Continua》 2025年第2期2495-2520,共26页
The widespread adoption of cloud computing has underscored the critical importance of efficient resource allocation and management, particularly in task scheduling, which involves assigning tasks to computing resource... The widespread adoption of cloud computing has underscored the critical importance of efficient resource allocation and management, particularly in task scheduling, which involves assigning tasks to computing resources for optimized resource utilization. Several meta-heuristic algorithms have shown effectiveness in task scheduling, among which the relatively recent Willow Catkin Optimization (WCO) algorithm has demonstrated potential, albeit with apparent needs for enhanced global search capability and convergence speed. To address these limitations of WCO in cloud computing task scheduling, this paper introduces an improved version termed the Advanced Willow Catkin Optimization (AWCO) algorithm. AWCO enhances the algorithm’s performance by augmenting its global search capability through a quasi-opposition-based learning strategy and accelerating its convergence speed via sinusoidal mapping. A comprehensive evaluation utilizing the CEC2014 benchmark suite, comprising 30 test functions, demonstrates that AWCO achieves superior optimization outcomes, surpassing conventional WCO and a range of established meta-heuristics. The proposed algorithm also considers trade-offs among the cost, makespan, and load balancing objectives. Experimental results of AWCO are compared with those obtained using the other meta-heuristics, illustrating that the proposed algorithm provides superior performance in task scheduling. The method offers a robust foundation for enhancing the utilization of cloud computing resources in the domain of task scheduling within a cloud computing environment. 展开更多
关键词 Willow catkin optimization algorithm cloud computing task scheduling opposition-based learning strategy
在线阅读 下载PDF
Multi-strategy Enhanced Hiking Optimization Algorithm for Task Scheduling in the Cloud Environment
6
作者 Libang Wu Shaobo Li +2 位作者 Fengbin Wu Rongxiang Xie Panliang Yuan 《Journal of Bionic Engineering》 2025年第3期1506-1534,共29页
Metaheuristic algorithms are pivotal in cloud task scheduling. However, the complexity and uncertainty of the scheduling problem severely limit algorithms. To bypass this circumvent, numerous algorithms have been prop... Metaheuristic algorithms are pivotal in cloud task scheduling. However, the complexity and uncertainty of the scheduling problem severely limit algorithms. To bypass this circumvent, numerous algorithms have been proposed. The Hiking Optimization Algorithm (HOA) have been used in multiple fields. However, HOA suffers from local optimization, slow convergence, and low efficiency of late iteration search when solving cloud task scheduling problems. Thus, this paper proposes an improved HOA called CMOHOA. It collaborates with multi-strategy to improve HOA. Specifically, Chebyshev chaos is introduced to increase population diversity. Then, a hybrid speed update strategy is designed to enhance convergence speed. Meanwhile, an adversarial learning strategy is introduced to enhance the search capability in the late iteration. Different scenarios of scheduling problems are used to test the CMOHOA’s performance. First, CMOHOA was used to solve basic cloud computing task scheduling problems, and the results showed that it reduced the average total cost by 10% or more. Secondly, CMOHOA has been applied to edge fog cloud scheduling problems, and the results show that it reduces the average total scheduling cost by 2% or more. Finally, CMOHOA reduced the average total cost by 7% or more in scheduling problems for information transmission. 展开更多
关键词 Task scheduling Chebyshev chaos Hybrid speed update strategy Metaheuristic algorithms The Hiking Optimization Algorithm(HOA)
在线阅读 下载PDF
Physical-layer secure hybrid task scheduling and resource management for fog computing IoT networks
7
作者 ZHANG Shibo GAO Hongyuan +1 位作者 SU Yumeng SUN Rongchen 《Journal of Systems Engineering and Electronics》 2025年第5期1146-1160,共15页
Fog computing has emerged as an important technology which can improve the performance of computation-intensive and latency-critical communication networks.Nevertheless,the fog computing Internet-of-Things(IoT)systems... Fog computing has emerged as an important technology which can improve the performance of computation-intensive and latency-critical communication networks.Nevertheless,the fog computing Internet-of-Things(IoT)systems are susceptible to malicious eavesdropping attacks during the information transmission,and this issue has not been adequately addressed.In this paper,we propose a physical-layer secure fog computing IoT system model,which is able to improve the physical layer security of fog computing IoT networks against the malicious eavesdropping of multiple eavesdroppers.The secrecy rate of the proposed model is analyzed,and the quantum galaxy–based search algorithm(QGSA)is proposed to solve the hybrid task scheduling and resource management problem of the network.The computational complexity and convergence of the proposed algorithm are analyzed.Simulation results validate the efficiency of the proposed model and reveal the influence of various environmental parameters on fog computing IoT networks.Moreover,the simulation results demonstrate that the proposed hybrid task scheduling and resource management scheme can effectively enhance secrecy performance across different communication scenarios. 展开更多
关键词 fog computing Internet-of-Things(IoT) physical layer security hybrid task scheduling and resource management quantum galaxy-based search algorithm(QGSA)
在线阅读 下载PDF
Improvised Seagull Optimization Algorithm for Scheduling Tasks in Heterogeneous Cloud Environment 被引量:2
8
作者 Pradeep Krishnadoss Vijayakumar Kedalu Poornachary +1 位作者 Parkavi Krishnamoorthy Leninisha Shanmugam 《Computers, Materials & Continua》 SCIE EI 2023年第2期2461-2478,共18页
Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis... Well organized datacentres with interconnected servers constitute the cloud computing infrastructure.User requests are submitted through an interface to these servers that provide service to them in an on-demand basis.The scientific applications that get executed at cloud by making use of the heterogeneous resources being allocated to them in a dynamic manner are grouped under NP hard problem category.Task scheduling in cloud poses numerous challenges impacting the cloud performance.If not handled properly,user satisfaction becomes questionable.More recently researchers had come up with meta-heuristic type of solutions for enriching the task scheduling activity in the cloud environment.The prime aim of task scheduling is to utilize the resources available in an optimal manner and reduce the time span of task execution.An improvised seagull optimization algorithm which combines the features of the Cuckoo search(CS)and seagull optimization algorithm(SOA)had been proposed in this work to enhance the performance of the scheduling activity inside the cloud computing environment.The proposed algorithm aims to minimize the cost and time parameters that are spent during task scheduling in the heterogeneous cloud environment.Performance evaluation of the proposed algorithm had been performed using the Cloudsim 3.0 toolkit by comparing it with Multi objective-Ant Colony Optimization(MO-ACO),ACO and Min-Min algorithms.The proposed SOA-CS technique had produced an improvement of 1.06%,4.2%,and 2.4%for makespan and had reduced the overall cost to the extent of 1.74%,3.93%and 2.77%when compared with PSO,ACO,IDEA algorithms respectively when 300 vms are considered.The comparative simulation results obtained had shown that the proposed improvised seagull optimization algorithm fares better than other contemporaries. 展开更多
关键词 Cloud computing task scheduling cuckoo search(CS) seagull optimization algorithm(SOA)
在线阅读 下载PDF
Dynamic access task scheduling of LEO constellation based on space-based distributed computing
9
作者 LIU Wei JIN Yifeng +2 位作者 ZHANG Lei GAO Zihe TAO Ying 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第4期842-854,共13页
A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process u... A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process under a series of complex constraints,which is important for enhancing the matching between resources and requirements.A complex algorithm is not available because that the LEO on-board resources is limi-ted.The proposed genetic algorithm(GA)based on two-dimen-sional individual model and uncorrelated single paternal inheri-tance method is designed to support distributed computation to enhance the feasibility of on-board application.A distributed system composed of eight embedded devices is built to verify the algorithm.A typical scenario is built in the system to evalu-ate the resource allocation process,algorithm mathematical model,trigger strategy,and distributed computation architec-ture.According to the simulation and measurement results,the proposed algorithm can provide an allocation result for more than 1500 tasks in 14 s and the success rate is more than 91%in a typical scene.The response time is decreased by 40%com-pared with the conditional GA. 展开更多
关键词 beam resource allocation distributed computing low Earth obbit(LEO)constellation spacecraft access task scheduling
在线阅读 下载PDF
Enhanced Hybrid Equilibrium Strategy in Fog-Cloud Computing Networks with Optimal Task Scheduling
10
作者 Muchang Rao Hang Qin 《Computers, Materials & Continua》 SCIE EI 2024年第5期2647-2672,共26页
More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud com... More devices in the Intelligent Internet of Things(AIoT)result in an increased number of tasks that require low latency and real-time responsiveness,leading to an increased demand for computational resources.Cloud computing’s low-latency performance issues in AIoT scenarios have led researchers to explore fog computing as a complementary extension.However,the effective allocation of resources for task execution within fog environments,characterized by limitations and heterogeneity in computational resources,remains a formidable challenge.To tackle this challenge,in this study,we integrate fog computing and cloud computing.We begin by establishing a fog-cloud environment framework,followed by the formulation of a mathematical model for task scheduling.Lastly,we introduce an enhanced hybrid Equilibrium Optimizer(EHEO)tailored for AIoT task scheduling.The overarching objective is to decrease both the makespan and energy consumption of the fog-cloud system while accounting for task deadlines.The proposed EHEO method undergoes a thorough evaluation against multiple benchmark algorithms,encompassing metrics likemakespan,total energy consumption,success rate,and average waiting time.Comprehensive experimental results unequivocally demonstrate the superior performance of EHEO across all assessed metrics.Notably,in the most favorable conditions,EHEO significantly diminishes both the makespan and energy consumption by approximately 50%and 35.5%,respectively,compared to the secondbest performing approach,which affirms its efficacy in advancing the efficiency of AIoT task scheduling within fog-cloud networks. 展开更多
关键词 Artificial intelligence of things fog computing task scheduling equilibrium optimizer differential evaluation algorithm local search
在线阅读 下载PDF
A Survey on Task Scheduling of CPU-GPU Heterogeneous Cluster
11
作者 ZHOU Yiheng ZENG Wei +2 位作者 ZHENG Qingfang LIU Zhilong CHEN Jianping 《ZTE Communications》 2024年第3期83-90,共8页
This paper reviews task scheduling frameworks,methods,and evaluation metrics of central processing unit-graphics processing unit(CPU-GPU)heterogeneous clusters.Task scheduling of CPU-GPU heterogeneous clusters can be ... This paper reviews task scheduling frameworks,methods,and evaluation metrics of central processing unit-graphics processing unit(CPU-GPU)heterogeneous clusters.Task scheduling of CPU-GPU heterogeneous clusters can be carried out on the system level,nodelevel,and device level.Most task-scheduling technologies are heuristic based on the experts’experience,while some technologies are based on statistic methods using machine learning,deep learning,or reinforcement learning.Many metrics have been adopted to evaluate and compare different task scheduling technologies that try to optimize different goals of task scheduling.Although statistic task scheduling has reached fewer research achievements than heuristic task scheduling,the statistic task scheduling still has significant research potential. 展开更多
关键词 CPU-GPU heterogeneous cluster task scheduling heuristic task scheduling statistic task scheduling PARALLELIZATION
在线阅读 下载PDF
Hybrid Task Scheduling Algorithm for Makespan Optimisation in Cloud Computing: A Performance Evaluation
12
作者 Abdulrahman M.Abdulghani 《Journal on Artificial Intelligence》 2024年第1期241-259,共19页
Cloud computing has rapidly evolved into a critical technology,seamlessly integrating into various aspects of daily life.As user demand for cloud services continues to surge,the need for efficient virtualization and r... Cloud computing has rapidly evolved into a critical technology,seamlessly integrating into various aspects of daily life.As user demand for cloud services continues to surge,the need for efficient virtualization and resource management becomes paramount.At the core of this efficiency lies task scheduling,a complex process that determines how tasks are allocated and executed across cloud resources.While extensive research has been conducted in the area of task scheduling,optimizing multiple objectives simultaneously remains a significant challenge due to the NP(Non-deterministic Polynomial)Complete nature of the problem.This study aims to address these challenges by providing a comprehensive review and experimental analysis of task scheduling approaches,with a particular focus on hybrid techniques that offer promising solutions.Utilizing the CloudSim simulation toolkit,we evaluated the performance of three hybrid algorithms:Estimation of Distribution Algorithm-Genetic Algorithm(EDA-GA),Hybrid Genetic Algorithm-Ant Colony Optimization(HGA-ACO),and Improved Discrete Particle Swarm Optimization(IDPSO).Our experimental results demonstrate that these hybrid methods significantly outperform traditional standalone algorithms in reducing Makespan,which is a critical measure of task completion time.Notably,the IDPSO algorithm exhibited superior performance,achieving a Makespan of just 0.64 milliseconds for a set of 150 tasks.These findings underscore the potential of hybrid algorithms to enhance task scheduling efficiency in cloud computing environments.This paper concludes with a discussion of the implications of our findings and offers recommendations for future research aimed at further improving task scheduling strategies,particularly in the context of increasingly complex and dynamic cloud environments. 展开更多
关键词 MAKESPAN multi-objective optimisation task scheduling cloud computing hybrid algorithms
在线阅读 下载PDF
Offload Strategy for Edge Computing in Satellite Networks Based on Software Defined Network 被引量:1
13
作者 Zhiguo Liu Yuqing Gui +1 位作者 Lin Wang Yingru Jiang 《Computers, Materials & Continua》 SCIE EI 2025年第1期863-879,共17页
Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in us... Satellite edge computing has garnered significant attention from researchers;however,processing a large volume of tasks within multi-node satellite networks still poses considerable challenges.The sharp increase in user demand for latency-sensitive tasks has inevitably led to offloading bottlenecks and insufficient computational capacity on individual satellite edge servers,making it necessary to implement effective task offloading scheduling to enhance user experience.In this paper,we propose a priority-based task scheduling strategy based on a Software-Defined Network(SDN)framework for satellite-terrestrial integrated networks,which clarifies the execution order of tasks based on their priority.Subsequently,we apply a Dueling-Double Deep Q-Network(DDQN)algorithm enhanced with prioritized experience replay to derive a computation offloading strategy,improving the experience replay mechanism within the Dueling-DDQN framework.Next,we utilize the Deep Deterministic Policy Gradient(DDPG)algorithm to determine the optimal resource allocation strategy to reduce the processing latency of sub-tasks.Simulation results demonstrate that the proposed d3-DDPG algorithm outperforms other approaches,effectively reducing task processing latency and thus improving user experience and system efficiency. 展开更多
关键词 Satellite network edge computing task scheduling computing offloading
在线阅读 下载PDF
ACS-based resource assignment and task scheduling in grid
14
作者 祁超 张璟 李军怀 《Journal of Southeast University(English Edition)》 EI CAS 2007年第3期451-454,共4页
To solve the deadlock problem of tasks that the interdependence between tasks fails to consider during the course of resource assignment and task scheduling based on the heuristics algorithm, an improved ant colony sy... To solve the deadlock problem of tasks that the interdependence between tasks fails to consider during the course of resource assignment and task scheduling based on the heuristics algorithm, an improved ant colony system (ACS) based algorithm is proposed. First, how to map the resource assignment and task scheduling (RATS) problem into the optimization selection problem of task resource assignment graph (TRAG) and to add the semaphore mechanism in the optimal TRAG to solve deadlocks are explained. Secondly, how to utilize the grid pheromone system model to realize the algorithm based on ACS is explicated. This refers to the construction of TRAG by the random selection of appropriate resources for each task by the user agent and the optimization of TRAG through the positive feedback and distributed parallel computing mechanism of the ACS. Simulation results show that the proposed algorithm is effective and efficient in solving the deadlock problem. 展开更多
关键词 GRID resource assignment task scheduling ant colony system (ACS) task resource assignment graph (TRAG) SEMAPHORE
在线阅读 下载PDF
Data Replica Location-Aware Joint Scheduling of Map and Reduce Tasks
15
作者 Yuqi Fan Bo Gao 《计算机科学与技术汇刊(中英文版)》 2025年第1期1-12,共12页
MapReduce is a popular data parallel processing framework in data centers.MapReduce splits a job into multiple map tasks and reduce tasks so that the tasks can be executed in parallel.Before running the map and reduce... MapReduce is a popular data parallel processing framework in data centers.MapReduce splits a job into multiple map tasks and reduce tasks so that the tasks can be executed in parallel.Before running the map and reduce tasks,the task nodes communicate with the data nodes to fetch the data required by the execution of the tasks.The network traffic between the nodes accounts for a big part of the running time of the MapReduce job.Therefore,careful map and reduce tasks scheduling is critical for MapReduce performance.Most of the current task scheduling algorithms only perform the scheduling for either map tasks or reduce tasks without the joint consideration of the impact of both map and reduce tasks scheduling on the network traffic.In this paper,we deal with the joint scheduling of map and reduce tasks problem with the aim to reduce the network traffic.We also propose a data replica Location-Aware Joint Scheduling of map and reduce tasks algorithm(LAJS).The algorithm determines the scheduling locations of map and reduce tasks according to the node processing capabilities and the data replica locations of the input data for the map tasks.We finally conduct experiments through simulations.Experiment results show that the proposed algorithm LAJS can effectively reduce the data traffic during job processing and improve job makespan performance. 展开更多
关键词 MAPREDUCE Task scheduling Data Location Network Traffic
在线阅读 下载PDF
An Effective Cloud Workflow Scheduling Approach Combining PSO and Idle Time Slot-Aware Rules 被引量:11
16
作者 Yun Wang Xingquan Zuo 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第5期1079-1094,共16页
Workflow scheduling is a key issue and remains a challenging problem in cloud computing.Faced with the large number of virtual machine(VM)types offered by cloud providers,cloud users need to choose the most appropriat... Workflow scheduling is a key issue and remains a challenging problem in cloud computing.Faced with the large number of virtual machine(VM)types offered by cloud providers,cloud users need to choose the most appropriate VM type for each task.Multiple task scheduling sequences exist in a workflow application.Different task scheduling sequences have a significant impact on the scheduling performance.It is not easy to determine the most appropriate set of VM types for tasks and the best task scheduling sequence.Besides,the idle time slots on VM instances should be used fully to increase resources'utilization and save the execution cost of a workflow.This paper considers these three aspects simultaneously and proposes a cloud workflow scheduling approach which combines particle swarm optimization(PSO)and idle time slot-aware rules,to minimize the execution cost of a workflow application under a deadline constraint.A new particle encoding is devised to represent the VM type required by each task and the scheduling sequence of tasks.An idle time slot-aware decoding procedure is proposed to decode a particle into a scheduling solution.To handle tasks'invalid priorities caused by the randomness of PSO,a repair method is used to repair those priorities to produce valid task scheduling sequences.The proposed approach is compared with state-of-the-art cloud workflow scheduling algorithms.Experiments show that the proposed approach outperforms the comparative algorithms in terms of both of the execution cost and the success rate in meeting the deadline. 展开更多
关键词 Cloud computing idle time slot particle swarm optimization task scheduling sequence workflow scheduling
在线阅读 下载PDF
A Parallel Genetic Simulated Annealing Hybrid Algorithm for Task Scheduling 被引量:12
17
作者 SHU Wanneng ZHENG Shijue 《Wuhan University Journal of Natural Sciences》 CAS 2006年第5期1378-1382,共5页
In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem i... In this paper combined with the advantages of genetic algorithm and simulated annealing, brings forward a parallel genetic simulated annealing hybrid algorithm (PGSAHA) and applied to solve task scheduling problem in grid computing. It first generates a new group of individuals through genetic operation such as reproduction, crossover, mutation, etc, and than simulated anneals independently all the generated individuals respectively. When the temperature in the process of cooling no longer falls, the result is the optimal solution on the whole. From the analysis and experiment result, it is concluded that this algorithm is superior to genetic algorithm and simulated annealing. 展开更多
关键词 grid computing task scheduling genetic algorithm simulated annealing PGSAHA algorithm
在线阅读 下载PDF
Task Scheduling for Multi-Cloud Computing Subject to Security and Reliability Constraints 被引量:8
18
作者 Qing-Hua Zhu Huan Tang +1 位作者 Jia-Jie Huang Yan Hou 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第4期848-865,共18页
The rise of multi-cloud systems has been spurred.For safety-critical missions,it is important to guarantee their security and reliability.To address trust constraints in a heterogeneous multi-cloud environment,this wo... The rise of multi-cloud systems has been spurred.For safety-critical missions,it is important to guarantee their security and reliability.To address trust constraints in a heterogeneous multi-cloud environment,this work proposes a novel scheduling method called matching and multi-round allocation(MMA)to optimize the makespan and total cost for all submitted tasks subject to security and reliability constraints.The method is divided into two phases for task scheduling.The first phase is to find the best matching candidate resources for the tasks to meet their preferential demands including performance,security,and reliability in a multi-cloud environment;the second one iteratively performs multiple rounds of re-allocating to optimize tasks execution time and cost by minimizing the variance of the estimated completion time.The proposed algorithm,the modified cuckoo search(MCS),hybrid chaotic particle search(HCPS),modified artificial bee colony(MABC),max-min,and min-min algorithms are implemented in CloudSim to create simulations.The simulations and experimental results show that our proposed method achieves shorter makespan,lower cost,higher resource utilization,and better trade-off between time and economic cost.It is more stable and efficient. 展开更多
关键词 Multi-cloud environment multi-quality of service(QoS) reliability SECURITY task scheduling
在线阅读 下载PDF
Fine-Grained Resource Provisioning and Task Scheduling for Heterogeneous Applications in Distributed Green Clouds 被引量:5
19
作者 Haitao Yuan Meng Chu Zhou +1 位作者 Qing Liu Abdullah Abusorrah 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2020年第5期1380-1393,共14页
An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years... An increasing number of enterprises have adopted cloud computing to manage their important business applications in distributed green cloud(DGC)systems for low response time and high cost-effectiveness in recent years.Task scheduling and resource allocation in DGCs have gained more attention in both academia and industry as they are costly to manage because of high energy consumption.Many factors in DGCs,e.g.,prices of power grid,and the amount of green energy express strong spatial variations.The dramatic increase of arriving tasks brings a big challenge to minimize the energy cost of a DGC provider in a market where above factors all possess spatial variations.This work adopts a G/G/1 queuing system to analyze the performance of servers in DGCs.Based on it,a single-objective constrained optimization problem is formulated and solved by a proposed simulated-annealing-based bees algorithm(SBA)to find SBA can minimize the energy cost of a DGC provider by optimally allocating tasks of heterogeneous applications among multiple DGCs,and specifying the running speed of each server and the number of powered-on servers in each GC while strictly meeting response time limits of tasks of all applications.Realistic databased experimental results prove that SBA achieves lower energy cost than several benchmark scheduling methods do. 展开更多
关键词 Bees algorithm data centers distributed green cloud(DGC) energy optimization intelligent optimization simulated annealing task scheduling machine learning
在线阅读 下载PDF
Variable scheduling interval task scheduling for phased array radar 被引量:5
20
作者 ZHANG Haowei XIE Junwei +2 位作者 ZHANG Zhaojian SHAO Lei CHEN Tangjun 《Journal of Systems Engineering and Electronics》 SCIE EI CSCD 2018年第5期937-946,共10页
A scheduling algorithm is presented aiming at the task scheduling problem in the phased array radar. Rather than assuming the scheduling interval(SI) time, which is the update interval of the radar invoking the schedu... A scheduling algorithm is presented aiming at the task scheduling problem in the phased array radar. Rather than assuming the scheduling interval(SI) time, which is the update interval of the radar invoking the scheduling algorithm, to be a fixed value,it is modeled as a fuzzy set to improve the scheduling flexibility.The scheduling algorithm exploits the fuzzy set model in order to intelligently adjust the SI time. The idle time in other SIs is provided for SIs which will be overload. Thereby more request tasks can be accommodated. The simulation results show that the proposed algorithm improves the successful scheduling ratio by 16%,the threat ratio of execution by 16% and the time utilization ratio by 15% compared with the highest task mode priority first(HPF)algorithm. 展开更多
关键词 phased array radar task scheduling variable scheduling interval(SI) fuzzy set
在线阅读 下载PDF
上一页 1 2 6 下一页 到第
使用帮助 返回顶部