期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Reinforcement Learning Based Quantization Strategy Optimal Assignment Algorithm for Mixed Precision
1
作者 Yuejiao Wang Zhong Ma +2 位作者 Chaojie Yang Yu Yang Lu Wei 《Computers, Materials & Continua》 SCIE EI 2024年第4期819-836,共18页
The quantization algorithm compresses the original network by reducing the numerical bit width of the model,which improves the computation speed. Because different layers have different redundancy and sensitivity to d... The quantization algorithm compresses the original network by reducing the numerical bit width of the model,which improves the computation speed. Because different layers have different redundancy and sensitivity to databit width. Reducing the data bit width will result in a loss of accuracy. Therefore, it is difficult to determinethe optimal bit width for different parts of the network with guaranteed accuracy. Mixed precision quantizationcan effectively reduce the amount of computation while keeping the model accuracy basically unchanged. In thispaper, a hardware-aware mixed precision quantization strategy optimal assignment algorithm adapted to low bitwidth is proposed, and reinforcement learning is used to automatically predict the mixed precision that meets theconstraints of hardware resources. In the state-space design, the standard deviation of weights is used to measurethe distribution difference of data, the execution speed feedback of simulated neural network accelerator inferenceis used as the environment to limit the action space of the agent, and the accuracy of the quantization model afterretraining is used as the reward function to guide the agent to carry out deep reinforcement learning training. Theexperimental results show that the proposed method obtains a suitable model layer-by-layer quantization strategyunder the condition that the computational resources are satisfied, and themodel accuracy is effectively improved.The proposed method has strong intelligence and certain universality and has strong application potential in thefield of mixed precision quantization and embedded neural network model deployment. 展开更多
关键词 Mixed precision quantization quantization strategy optimal assignment reinforcement learning neural network model deployment
在线阅读 下载PDF
Optimization of Quantizer’s Segment Threshold Using Spline Approximations for Optimal Compressor Function
2
作者 Lazar Velimirovic Zoran Peric +1 位作者 Miomir Stankovic Jelena Nikolic 《Applied Mathematics》 2012年第10期1430-1434,共5页
In this paper, the optimization of quantizer’s segment threshold is done. The quantizer is designed on the basis of approximative spline functions. Coefficients on which we form approximative spline functions are cal... In this paper, the optimization of quantizer’s segment threshold is done. The quantizer is designed on the basis of approximative spline functions. Coefficients on which we form approximative spline functions are calculated by minimization mean square error (MSE). For coefficients determined in this way, spline functions by which optimal compressor function is approximated are obtained. For the quantizer designed on the basis of approximative spline functions, segment threshold is numerically determined depending on maximal value of the signal to quantization noise ratio (SQNR). Thus, quantizer with optimized segment threshold is achieved. It is shown that by quantizer model designed in this way and proposed in this paper, the SQNR that is very close to SQNR of nonlinear optimal companding quantizer is achieved. 展开更多
关键词 optimization of Quantizer’s Segment Threshold Mean Square Error Second-Degree Spline Functions Compressor Function
在线阅读 下载PDF
Training time minimization for federated edge learning with optimized gradient quantization and bandwidth allocation
3
作者 Peixi LIU Jiamo JIANG +5 位作者 Guangxu ZHU Lei CHENG Wei JIANG Wu LUO Ying DU Zhiqin WANG 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2022年第8期1247-1263,共17页
Training a machine learning model with federated edge learning(FEEL)is typically time consuming due to the constrained computation power of edge devices and the limited wireless resources in edge networks.In this stud... Training a machine learning model with federated edge learning(FEEL)is typically time consuming due to the constrained computation power of edge devices and the limited wireless resources in edge networks.In this study,the training time minimization problem is investigated in a quantized FEEL system,where heterogeneous edge devices send quantized gradients to the edge server via orthogonal channels.In particular,a stochastic quantization scheme is adopted for compression of uploaded gradients,which can reduce the burden of per-round communication but may come at the cost of increasing the number of communication rounds.The training time is modeled by taking into account the communication time,computation time,and the number of communication rounds.Based on the proposed training time model,the intrinsic trade-off between the number of communication rounds and per-round latency is characterized.Specifically,we analyze the convergence behavior of the quantized FEEL in terms of the optimality gap.Furthermore,a joint data-and-model-driven fitting method is proposed to obtain the exact optimality gap,based on which the closed-form expressions for the number of communication rounds and the total training time are obtained.Constrained by the total bandwidth,the training time minimization problem is formulated as a joint quantization level and bandwidth allocation optimization problem.To this end,an algorithm based on alternating optimization is proposed,which alternatively solves the subproblem of quantization optimization through successive convex approximation and the subproblem of bandwidth allocation by bisection search.With different learning tasks and models,the validation of our analysis and the near-optimal performance of the proposed optimization algorithm are demonstrated by the simulation results. 展开更多
关键词 Federated edge learning quantization optimization Bandwith allocation Training time minimization
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部