期刊文献+
共找到528篇文章
< 1 2 27 >
每页显示 20 50 100
Dynamic access task scheduling of LEO constellation based on space-based distributed computing 被引量:2
1
作者 LIU Wei JIN Yifeng +2 位作者 ZHANG Lei GAO Zihe TAO Ying 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第4期842-854,共13页
A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process u... A dynamic multi-beam resource allocation algorithm for large low Earth orbit(LEO)constellation based on on-board distributed computing is proposed in this paper.The allocation is a combinatorial optimization process under a series of complex constraints,which is important for enhancing the matching between resources and requirements.A complex algorithm is not available because that the LEO on-board resources is limi-ted.The proposed genetic algorithm(GA)based on two-dimen-sional individual model and uncorrelated single paternal inheri-tance method is designed to support distributed computation to enhance the feasibility of on-board application.A distributed system composed of eight embedded devices is built to verify the algorithm.A typical scenario is built in the system to evalu-ate the resource allocation process,algorithm mathematical model,trigger strategy,and distributed computation architec-ture.According to the simulation and measurement results,the proposed algorithm can provide an allocation result for more than 1500 tasks in 14 s and the success rate is more than 91%in a typical scene.The response time is decreased by 40%com-pared with the conditional GA. 展开更多
关键词 beam resource allocation distributed computing low Earth obbit(LEO)constellation spacecraft access task scheduling
在线阅读 下载PDF
CPTF–a new heuristic based branch and bound algorithm for workflow scheduling in heterogeneous distributed computing systems
2
作者 D.Sirisha S.Sambhu Prasad 《CCF Transactions on High Performance Computing》 2024年第5期472-487,共16页
Computationally intensive applications embodied as workflows entail interdependent tasks that involve multifarious computation requirements and necessitate Heterogeneous Distributed Computing Systems(HDCS)to attain hi... Computationally intensive applications embodied as workflows entail interdependent tasks that involve multifarious computation requirements and necessitate Heterogeneous Distributed Computing Systems(HDCS)to attain high performance.The scheduling of workflows on HDCS was demonstrated as an NP-Complete problem.In the current work,a new heuristic based Branch and Bound(BnB)technique namely Critical Path_finish Time First(CPTF)algorithm is proposed for workflow scheduling on HDCS to achieve the best solutions.The primary merits of CPTF algorithm are due to the bounding functions that are tight and of less complexity.The sharp bounding functions could precisely estimate the promise of each state and aid in pruning infeasible states.Thus,the search space size is reduced.The CPTF algorithm explores the most promising states in the search space and converges to the solution quickly.Therefore,high performance is achieved.The experimental results on random and scientific workflows reveal that CPTF algorithm could effectively exploit high potency of BnB technique in realizing better quality solutions against the widely referred heuristic scheduling algorithms.The results on the benchmark workflows show that CPTF algorithm has improved schedules for 89.36%of the cases. 展开更多
关键词 Workflow scheduling Task scheduling Heuristics Heterogeneous distributed computing systems Branch and bound technique Makespan
在线阅读 下载PDF
FedCCM:Communication-Efficient Federated Learning via Clustered Client Momentum in Non-IID Settings
3
作者 Hang Wen Kai Zeng 《Computers, Materials & Continua》 2026年第3期1690-1707,共18页
Federated learning often experiences slow and unstable convergence due to edge-side data heterogeneity.This problem becomes more severe when edge participation rate is low,as the information collected from different e... Federated learning often experiences slow and unstable convergence due to edge-side data heterogeneity.This problem becomes more severe when edge participation rate is low,as the information collected from different edge devices varies significantly.As a result,communication overhead increases,which further slows down the convergence process.To address this challenge,we propose a simple yet effective federated learning framework that improves consistency among edge devices.The core idea is clusters the lookahead gradients collected from edge devices on the cloud server to obtain personalized momentum for steering local updates.In parallel,a global momentum is applied during model aggregation,enabling faster convergence while preserving personalization.This strategy enables efficient propagation of the estimated global update direction to all participating edge devices and maintains alignment in local training,without introducing extra memory or communication overhead.We conduct extensive experiments on benchmark datasets such as Cifar100 and Tiny-ImageNet.The results confirm the effectiveness of our framework.On CIFAR-100,our method reaches 55%accuracy with 37 fewer rounds and achieves a competitive final accuracy of 65.46%.Even under extreme non-IID scenarios,it delivers significant improvements in both accuracy and communication efficiency.The implementation is publicly available at https://github.com/sjmp525/CollaborativeComputing/tree/FedCCM(accessed on 20 October 2025). 展开更多
关键词 Federated learning distributed computation communication efficient momentum clustering non-independent and identically distributed(non-IID)
在线阅读 下载PDF
A Multi-Objective Clustered Input Oriented Salp Swarm Algorithm in Cloud Computing
4
作者 Juliet A.Murali Brindha T. 《Computers, Materials & Continua》 SCIE EI 2024年第12期4659-4690,共32页
Infrastructure as a Service(IaaS)in cloud computing enables flexible resource distribution over the Internet,but achieving optimal scheduling remains a challenge.Effective resource allocation in cloud-based environmen... Infrastructure as a Service(IaaS)in cloud computing enables flexible resource distribution over the Internet,but achieving optimal scheduling remains a challenge.Effective resource allocation in cloud-based environments,particularly within the IaaS model,poses persistent challenges.Existing methods often struggle with slow opti-mization,imbalanced workload distribution,and inefficient use of available assets.These limitations result in longer processing times,increased operational expenses,and inadequate resource deployment,particularly under fluctuating demands.To overcome these issues,a novel Clustered Input-Oriented Salp Swarm Algorithm(CIOSSA)is introduced.This approach combines two distinct strategies:Task Splitting Agglomerative Clustering(TSAC)with an Input Oriented Salp Swarm Algorithm(IOSSA),which prioritizes tasks based on urgency,and a refined multi-leader model that accelerates optimization processes,enhancing both speed and accuracy.By continuously assessing system capacity before task distribution,the model ensures that assets are deployed effectively and costs are controlled.The dual-leader technique expands the potential solution space,leading to substantial gains in processing speed,cost-effectiveness,asset efficiency,and system throughput,as demonstrated by comprehensive tests.As a result,the suggested model performs better than existing approaches in terms of makespan,resource utilisation,throughput,and convergence speed,demonstrating that CIOSSA is scalable,reliable,and appropriate for the dynamic settings found in cloud computing. 展开更多
关键词 Cloud computing clustering resource allocation scheduling swam algorithms optimization common with in the subject discipline
在线阅读 下载PDF
Extended Balanced Scheduler with Clustering and Replication for Data Intensive Scientific Workflow Applications in Cloud Computing
5
作者 Satwinder Kaur Mehak Aggarwal 《Journal of Electronic Research and Application》 2018年第3期8-15,共8页
Cloud computing is an advance computing model using which several applications,data and countless IT services are provided over the Internet.Task scheduling plays a crucial role in cloud computing systems.The issue of... Cloud computing is an advance computing model using which several applications,data and countless IT services are provided over the Internet.Task scheduling plays a crucial role in cloud computing systems.The issue of task scheduling can be viewed as the finding or searching an optimal mapping/assignment of set of subtasks of different tasks over the available set of resources so that we can achieve the desired goals for tasks.With the enlargement of users of cloud the tasks need to be scheduled.Cloud’s performance depends on the task scheduling algorithms used.Numerous algorithms have been submitted in the past to solve the task scheduling problem for heterogeneous network of computers.The existing research work proposes different methods for data intensive applications which are energy and deadline aware task scheduling method.As scientific workflow is combination of fine grain and coarse grain task.Every task scheduled to VM has system overhead.If multiple fine grain task are executing in scientific workflow,it increase the scheduling overhead.To overcome the scheduling overhead,multiple small tasks has been combined to large task,which decrease the scheduling overhead and improve the execution time of the workflow.Horizontal clustering has been used to cluster the fine grained task further replication technique has been combined.The proposed scheduling algorithm improves the performance metrics such as execution time and cost.Further this research can be extended with improved clustering technique and replication methods. 展开更多
关键词 SCIENTIFIC WORKFLOW cloud computing REPLICATION clusterING scheduling
在线阅读 下载PDF
Cluster-Based Distributed Algorithms for Very Large Linear Equations
6
作者 古志民 MARTA Kwiatkowska 付引霞 《Journal of Beijing Institute of Technology》 EI CAS 2006年第1期66-70,共5页
In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot ... In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot solve the problems. A distributed cluster-based solution for very large linear equations is discussed, it includes the definitions of notations, partition of matrix, communication mechanism, and a master-slaver algorithm etc., the computing cost is O(n^3/N), the memory cost is O(n^2/N), the I/O cost is O(n^2/N), and the com- munication cost is O(Nn ), here, N is the number of computing nodes or processes. Some tests show that the solution could solve the double type of matrix under 10^6 × 10^6 effectively. 展开更多
关键词 Gaussian elimination PARTITION cluster-based distributed computing
在线阅读 下载PDF
Trusted Data Acquisition Mechanism for Cloud Resource Scheduling Based on Distributed Agents 被引量:4
7
作者 李小勇 杨月华 《China Communications》 SCIE CSCD 2011年第6期108-116,共9页
Goud computing is a new paradigm in which dynamic and virtualized computing resources are provided as services over the Internet. However, because cloud resource is open and dynamically configured, resource allocation... Goud computing is a new paradigm in which dynamic and virtualized computing resources are provided as services over the Internet. However, because cloud resource is open and dynamically configured, resource allocation and scheduling are extremely important challenges in cloud infrastructure. Based on distributed agents, this paper presents trusted data acquisition mechanism for efficient scheduling cloud resources to satisfy various user requests. Our mechanism defines, collects and analyzes multiple key trust targets of cloud service resources based on historical information of servers in a cloud data center. As a result, using our trust computing mechanism, cloud providers can utilize their resources efficiently and also provide highly trusted resources and services to many users. 展开更多
关键词 cloud computing trusted computing distributed agent resource scheduling
在线阅读 下载PDF
A Survey of Spark Scheduling Strategy Optimization Techniques and Development Trends
8
作者 Chuan Li Xuanlin Wen 《Computers, Materials & Continua》 2025年第6期3843-3875,共33页
Spark performs excellently in large-scale data-parallel computing and iterative processing.However,with the increase in data size and program complexity,the default scheduling strategy has difficultymeeting the demand... Spark performs excellently in large-scale data-parallel computing and iterative processing.However,with the increase in data size and program complexity,the default scheduling strategy has difficultymeeting the demands of resource utilization and performance optimization.Scheduling strategy optimization,as a key direction for improving Spark’s execution efficiency,has attracted widespread attention.This paper first introduces the basic theories of Spark,compares several default scheduling strategies,and discusses common scheduling performance evaluation indicators and factors affecting scheduling efficiency.Subsequently,existing scheduling optimization schemes are summarized based on three scheduling modes:load characteristics,cluster characteristics,and matching of both,and representative algorithms are analyzed in terms of performance indicators and applicable scenarios,comparing the advantages and disadvantages of different scheduling modes.The article also explores in detail the integration of Spark scheduling strategies with specific application scenarios and the challenges in production environments.Finally,the limitations of the existing schemes are analyzed,and prospects are envisioned. 展开更多
关键词 SPARK scheduling optimization load balancing resource utilization distributed computing
在线阅读 下载PDF
An Optimized Resource Scheduling Strategy for Hadoop Speculative Execution Based on Non-cooperative Game Schemes
9
作者 Yinghang Jiang Qi Liu +3 位作者 Williams Dannah Dandan Jin Xiaodong Liu Mingxu Sun 《Computers, Materials & Continua》 SCIE EI 2020年第2期713-729,共17页
Hadoop is a well-known parallel computing system for distributed computing and large-scale data processes.“Straggling”tasks,however,have a serious impact on task allocation and scheduling in a Hadoop system.Speculat... Hadoop is a well-known parallel computing system for distributed computing and large-scale data processes.“Straggling”tasks,however,have a serious impact on task allocation and scheduling in a Hadoop system.Speculative Execution(SE)is an efficient method of processing“Straggling”Tasks by monitoring real-time running status of tasks and then selectively backing up“Stragglers”in another node to increase the chance to complete the entire mission early.Present speculative execution strategies meet challenges on misjudgement of“Straggling”tasks and improper selection of backup nodes,which leads to inefficient implementation of speculative executive processes.This paper has proposed an Optimized Resource Scheduling strategy for Speculative Execution(ORSE)by introducing non-cooperative game schemes.The ORSE transforms the resource scheduling of backup tasks into a multi-party non-cooperative game problem,where the tasks are regarded as game participants,whilst total task execution time of the entire cluster as the utility function.In that case,the most benefit strategy can be implemented in each computing node when the game reaches a Nash equilibrium point,i.e.,the final resource scheduling scheme to be obtained.The strategy has been implemented in Hadoop-2.x.Experimental results depict that the ORSE can maintain the efficiency of speculative executive processes and improve fault-tolerant and computation performance under the circumstances of Normal Load,Busy Load and Busy Load with Skewed Data. 展开更多
关键词 distributed computing speculative execution resource scheduling non-cooperative game theory
在线阅读 下载PDF
Method for improving MapReduce performance by prefetching before scheduling
10
作者 张霄宏 Feng Shengzhong +1 位作者 Fan Jianping Huang Zhexue 《High Technology Letters》 EI CAS 2012年第4期343-349,共7页
In this paper, a prefetching technique is proposed to solve the performance problem caused by remote data access delay. In the technique, the map tasks which will cause the delay are predicted first and then the input... In this paper, a prefetching technique is proposed to solve the performance problem caused by remote data access delay. In the technique, the map tasks which will cause the delay are predicted first and then the input data of these tasks will be preloaded before the tasks are scheduled. During the execution, the input data can be read from local nodes. Therefore, the delay can be hidden. The technique has been implemented in Hadoop-0. 20.1. The experiment results have shown that the technique reduces map tasks causing delay, and improves the performance of Hadoop MapRe- duce by 20%. 展开更多
关键词 cloud computing distributed computing PREFETCHING MAPREDUCE scheduling
在线阅读 下载PDF
Self Organization Map for Clustering and Classification in the Ecology of Agent Organizations 被引量:3
11
作者 Dimuthu Chandana Kelegama LIU Li-hua LIU Jian-qin 《Journal of Central South University》 SCIE EI CAS 2000年第1期53-56,共4页
Development of computational agent organizations or “societies” has become the domiant computing paradigm in the arena of Distributed Artificial Intelligence, and many foreseeable future applications need agent orga... Development of computational agent organizations or “societies” has become the domiant computing paradigm in the arena of Distributed Artificial Intelligence, and many foreseeable future applications need agent organizations, in which diversified agents cooperate in a distributed manner, forming teams. In such scenarios, the agents would need to know each other in order to facilitate the interactions. Moreover, agents in such an environment are not statically defined in advance but they can adaptively enter and leave an organization. This begs the question of how agents locate each other in order to cooperate in achieving organizational goals. Locating agents is a quite challenging task, especially in organizations that involve a large number of agents and where the resource avaiability is intermittent. The authors explore here an approach based on self organization map (SOM) which will serve as a clustering method in the light of the knowledge gathered about various agents. The approach begins by categorizing agents using a selected set of agent properties. These categories are used to derive various ranks and a distance matrix. The SOM algorithm uses this matrix as input to obtain clusters of agents. These clusters reduce the search space, resulting in a relatively short agent search time. 展开更多
关键词 clusterING classification AGENT organizations AGENT societies self ORGANIZING distributed computing
在线阅读 下载PDF
PRI: An Periodically Receiver-Initiated Task Scheduling Algorithm
12
作者 石威 《High Technology Letters》 EI CAS 2000年第1期10-15,共6页
Task scheduling is a key problem for the distributed computation. This thesis analyzes receiver initiated(RI) task scheduling algorithm, finds its weakness and presents an improved algorithm PRI algorithm. This algo... Task scheduling is a key problem for the distributed computation. This thesis analyzes receiver initiated(RI) task scheduling algorithm, finds its weakness and presents an improved algorithm PRI algorithm. This algorithm schedules the concurrent tasks onto network of workstation dynamically at runtime, and initiates task scheduling by the node of low load. The threshold on each node can be modified according to the system information which is periodically detected. Meanwhile, the detecting period can be adjusted in terms of the change of the system state. The result of the experiments shows that the PRI algorithm is superior to the RI algorithm. 展开更多
关键词 Task scheduling distributed computation RECEIVER initiated Network of WORKSTATIONS RUNTIME Low load THRESHOLD
在线阅读 下载PDF
DSparse:A Distributed Training Method for Edge Clusters Based on Sparse Update 被引量:1
13
作者 Xiao-Hui Peng Yi-Xuan Sun +1 位作者 Zheng-Hui Zhang Yi-Fan Wang 《Journal of Computer Science & Technology》 2025年第3期637-653,共17页
Edge machine learning creates a new computational paradigm by enabling the deployment of intelligent applications at the network edge.It enhances application efficiency and responsiveness by performing inference and t... Edge machine learning creates a new computational paradigm by enabling the deployment of intelligent applications at the network edge.It enhances application efficiency and responsiveness by performing inference and training tasks closer to data sources.However,it encounters several challenges in practice.The variance in hardware specifications and performance across different devices presents a major issue for the training and inference tasks.Additionally,edge devices typically possess limited network bandwidth and computing resources compared with data centers.Moreover,existing distributed training architectures often fail to consider the constraints of resources and communication efficiency in edge environments.In this paper,we propose DSparse,a method for distributed training based on sparse update in edge clusters with various memory capacities.It aims at maximizing the utilization of memory resources across all devices within a cluster.To reduce memory consumption during the training process,we adopt sparse update to prioritize the updating of selected layers on the devices in the cluster,which not only lowers memory usage but also reduces the data volume of parameters and the time required for parameter aggregation.Furthermore,DSparse utilizes a parameter aggregation mechanism based on multi-process groups,subdividing the aggregation tasks into AllReduce and Broadcast types,thereby further reducing the communication frequency for parameter aggregation.Experimental results using the MobileNetV2 model on the CIFAR-10 dataset demonstrate that DSparse reduces memory consumption by an average of 59.6%across seven devices,with a 75.4%reduction in parameter aggregation time,while maintaining model precision. 展开更多
关键词 distributed training edge computing edge machine learning sparse update edge cluster
原文传递
Optimal Scheduling of Distribution System with Edge Computing and Data-driven Modeling of Demand Response 被引量:2
14
作者 Jianpei Han Nian Liu Jiaqi Shi 《Journal of Modern Power Systems and Clean Energy》 SCIE EI CSCD 2022年第4期989-999,共11页
High penetration of renewable energies enlarge the peak-valley difference of the net load of the distribution system,which puts forward higher requirements for the operation scheduling of the distribution system.From ... High penetration of renewable energies enlarge the peak-valley difference of the net load of the distribution system,which puts forward higher requirements for the operation scheduling of the distribution system.From the perspective of leveraging demand-side adjustment capabilities,an optimal scheduling method of the distribution system with edge computing and data-driven modeling of price-based demand response(PBDR)is proposed.By introducing the edge computing paradigm,a collaborative interaction framework between the control center and the edge nodes is designed for the optimization of the distribution system.At the edge nodes,a classified XGBoost-based PBDR modeling method is proposed for large-scale differentiated users.At the control center,a two-stage optimization method integrating pre-scheduling and re-scheduling is proposed based on demand response results from all edge nodes.Through the information interaction between the control center and edge nodes,the optimized scheduling of the distribution system with large-scale users is realized.Finally,a case study is implemented on the modified IEEE 33-node system,which verifies that the proposed classified modeling method has lower errors,and it is beneficial to improve the economics of the system operation.Moreover,the simulation results show that the application of edge computing can significantly reduce the calculation time of the optimal scheduling problem with PBDR modeling of large-scale users. 展开更多
关键词 Demand response distribution system edge computing optimal scheduling XGBoost
原文传递
On/Off-Line Prediction Applied to Job Scheduling on Non-Dedicated NOWs 被引量:1
15
作者 Mauricio Hanzich Porfidio Hernandez +2 位作者 Francesc Gine Francesc Solsona Josep L.Lerida 《Journal of Computer Science & Technology》 SCIE EI CSCD 2011年第1期99-116,共18页
This paper proposes a prediction engine designed for non-dedicated clusters, which is able to estimate the turnaround time for parallel applications, even in the presence of serial workload of the workstation owner. T... This paper proposes a prediction engine designed for non-dedicated clusters, which is able to estimate the turnaround time for parallel applications, even in the presence of serial workload of the workstation owner. The prediction engine can be configured to work with three different estimation kernels: a Historical kernel, a Simulation kernel based on analytical models and an integration of both, named Hybrid kernel. These estimation proposals were integrated into a scheduling system, named CISNE, which can be executed in an on-line or off-line mode. The accuracy of the proposed estimation methods was evaluated in relation to different job scheduling policies in a real and a simulated cluster environment. In both environments, we observed that the Hybrid system gives the best results because it combines the ability of a simulation engine to capture the dynamism of a non-dedicated environment together with the accuracy of the historical methods to estimate the application runtime considering the state of the resources. 展开更多
关键词 prediction method non-dedicated cluster cluster computing job scheduling simulation
原文传递
AI-Geo:基于并行分布式架构的智能地球物理云平台
16
作者 陈鸿龙 马博仑 +2 位作者 李哲 高志敏 宋建国 《计算机应用与软件》 北大核心 2026年第1期9-16,共8页
针对物探行业存在的大数据处理缓慢、软件繁琐、算法难以应用等问题,设计开发出一套基于并行分布式架构的云平台。该平台通过Web使用,采用模块化与工作流交互,具备并行计算与分布式系统。平台部署地学常用算法与AI框架,支持模块化扩展,... 针对物探行业存在的大数据处理缓慢、软件繁琐、算法难以应用等问题,设计开发出一套基于并行分布式架构的云平台。该平台通过Web使用,采用模块化与工作流交互,具备并行计算与分布式系统。平台部署地学常用算法与AI框架,支持模块化扩展,并具备多语言与多环境聚合系统,兼容不同平台算法。平台具备知识共享和资料管理系统,并能自适应切换多种数据源。实验表明,平台能显著缩短运行耗时,并可面向物探信息解释的多元化场景,促进行业高质量发展。 展开更多
关键词 分布式集群 并行计算 云计算 工作流 地球物理
在线阅读 下载PDF
面向石油物探领域的HPC故障与性能分析
17
作者 王明键 赵长海 +3 位作者 李超 文佳敏 尚民强 侯红军 《北京航空航天大学学报》 北大核心 2026年第1期157-166,共10页
如今,石油物探领域地震资料采集技术不断进步,数据规模已达到TB甚至PB级别,随着数据量,运行时间,以及高性能计算(HPC)集群中节点数的增加,集群出现问题的概率和维护难度也随之增加。当集群或节点出现故障时往往需要重新运行计算程序,造... 如今,石油物探领域地震资料采集技术不断进步,数据规模已达到TB甚至PB级别,随着数据量,运行时间,以及高性能计算(HPC)集群中节点数的增加,集群出现问题的概率和维护难度也随之增加。当集群或节点出现故障时往往需要重新运行计算程序,造成了极大的资源浪费。为解决计算集群中HPC程序可观测性低、故障和性能分析困难的问题,借鉴开放追踪标准(OTF)与分布式链路追踪的思想,提出一种面向生产环境的低侵入式高性能计算集群和程序故障分析方法,该方法不仅能够对生产环境中HPC程序进行高效观测,还具有低侵入性的特点,可以在几乎不修改代码的前提下与现有应用程序结合使用。将所提方法用于生产环境中分布式“抽道集”排序程序进行采集分析,验证了所提方法的有效性,发现了程序中隐藏的软件缺陷和性能薄弱环节。 展开更多
关键词 分布式计算 大规模集群 故障分析 链路追踪 生产环境
原文传递
Consolidated cluster systems for data centers in the cloud age: a survey and analysis 被引量:2
18
作者 Jian LIN Li ZHA Zhiwei XU 《Frontiers of Computer Science》 SCIE EI CSCD 2013年第1期1-19,共19页
In the cloud age, heterogeneous application modes on large-scale infrastructures bring about the chal- lenges on resource utilization and manageability to data cen- ters. Many resource and runtime management systems a... In the cloud age, heterogeneous application modes on large-scale infrastructures bring about the chal- lenges on resource utilization and manageability to data cen- ters. Many resource and runtime management systems are developed or evolved to address these challenges and rele- vant problems from different perspectives. This paper tries to identify the main motivations, key concerns, common fea- tures, and representative solutions of such systems through a survey and analysis. A typical kind of these systems is gener- alized as the consolidated cluster system, whose design goal is identified as reducing the overall costs under the quality of service premise. A survey on this kind of systems is given, and the critical issues concerned by such systems are sum- marized as resource consolidation and runtime coordination. These two issues are analyzed and classified according to the design styles and external characteristics abstracted from the surveyed work. Five representative consolidated cluster systems from both academia and industry are illustrated and compared in detail based on the analysis and classifications. We hope this survey and analysis to be conducive to both de- sign implementation and technology selection of this kind of systems, in response to the constantly emerging challenges on infrastructure and application management in data centers. 展开更多
关键词 data center cloud computing distributed re- source management consolidated cluster system resource consolidation runtime coordination
原文传递
基于边缘计算与深度强化学习的主动配电网实时优化调度策略
19
作者 李武 高奇 +4 位作者 杨慧 闫凯文 阮玉园 赵英男 刘俊 《电测与仪表》 北大核心 2026年第1期105-114,共10页
主动配电网中新能源渗透比例的不断增加,导致运行数据激增,而新能源的间歇特性等因素也对调度策略产生了重大挑战。基于此,文中提出VMD-BiLSTM-PPO的实时优化调度模型。该模型基于边缘计算,构建多区域能源自治框架,采用深度强化学习的... 主动配电网中新能源渗透比例的不断增加,导致运行数据激增,而新能源的间歇特性等因素也对调度策略产生了重大挑战。基于此,文中提出VMD-BiLSTM-PPO的实时优化调度模型。该模型基于边缘计算,构建多区域能源自治框架,采用深度强化学习的近端策略优化(proximal policy optimization,PPO)算法,以运行调度成本最小为目标,实现配电网云-边协同的优化调度。该模型将大量计算和数据存储任务下放至边缘侧,可以有效减少调度中心的计算量和数据传输量。在PPO算法中,采用基于变分模态分解(variational mode decomposition,VMD)和双向长短期记忆网络(Bi-directional long short-term memory,BiLSTM)的新能源出力预测,可以有效缓解新能源波动性带来的影响。仿真实验结果表明该模型能够提高新能源的消纳率,并提升配电网实时调度的经济性。 展开更多
关键词 主动配电网 实时调度 边缘计算 深度强化学习
在线阅读 下载PDF
面向异构AI负载的动态融合型智能算力集群架构设计
20
作者 赵伟 毛磊 +1 位作者 邱瑞文 周宏成 《中国建设信息化》 2026年第4期74-78,共5页
基于传统算力集群无法适应异构硬件环境下资源高效利用的现实问题,通过研究异构AI负载、异构算力特征及现有动态资源管理技术,给出了一种面向异构AI负载的动态融合型智能算力集群架构。在此基础上,提出了一种基于深度强化学习的任务感... 基于传统算力集群无法适应异构硬件环境下资源高效利用的现实问题,通过研究异构AI负载、异构算力特征及现有动态资源管理技术,给出了一种面向异构AI负载的动态融合型智能算力集群架构。在此基础上,提出了一种基于深度强化学习的任务感知智能调度算法,对实现集群任务调度和资源分配具有创新的理论指导意义。 展开更多
关键词 异构AI负载 任务调度 算力集群 资源分配
在线阅读 下载PDF
上一页 1 2 27 下一页 到第
使用帮助 返回顶部