期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
vTensor-Based GPU Memory Management for Edge Deep Learning Training
1
作者 he-ran gao Diao-Han Luo +2 位作者 Yue-Wen Wu Heng Wu Wen-Bo Zhang 《Journal of Computer Science & Technology》 2025年第6期1608-1625,共18页
Supporting real-time and privacy-preserving learning at the edge is emerging as a critical trend,bringing forth substantial challenges for deep learning(DL)training in the context of limited GPU(graphic processing uni... Supporting real-time and privacy-preserving learning at the edge is emerging as a critical trend,bringing forth substantial challenges for deep learning(DL)training in the context of limited GPU(graphic processing unit)memory.Recent work has sought to address the limitations by swapping tensors between GPU memory and CPU memory.Unfortunately,their tensor-based memory management encounters additional overhead since the swapped tensors do not align with the actual memory demands,resulting in decreased throughput.This paper introduces a vTensor-based memory management approach designed to mitigate memory swapping overhead.Virtualized tensors,dubbed vTensors,are used to finely align memory swapping amounts with real-time memory demands.Firstly,we introduce an abstraction layer that virtualizes coarse-grained tensors to multiple finer-grained vTensors.Secondly,we propose the Layered Graph Model(LGM)for analyzing vTensor mappings,which produces a memory swapping plan leveraged in the subsequent DL training iterations.Evaluations conducted on typical edge deep learning models illustrate that our approach surpasses prior work with a 15.60%increase in DL training throughput. 展开更多
关键词 deep learning training vTensor memory management memory swapping memory layout
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部